[ 468.155070] env[61243]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61243) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 468.155461] env[61243]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61243) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 468.155461] env[61243]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61243) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 468.155788] env[61243]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 468.248215] env[61243]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61243) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 468.257889] env[61243]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61243) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 468.858734] env[61243]: INFO nova.virt.driver [None req-f884f6a0-eb96-431b-9d32-d08740e4a07a None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 468.932106] env[61243]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 468.932283] env[61243]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 468.932381] env[61243]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61243) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 472.202268] env[61243]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-cfaf6522-555a-4a3a-8d68-373d83892b4d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.233124] env[61243]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61243) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 472.233390] env[61243]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-8e9f700b-8eae-42ae-bc6a-70307bc909c3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.274554] env[61243]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 1e084. [ 472.274784] env[61243]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.342s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 472.275319] env[61243]: INFO nova.virt.vmwareapi.driver [None req-f884f6a0-eb96-431b-9d32-d08740e4a07a None None] VMware vCenter version: 7.0.3 [ 472.278758] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88fc520-701b-4e02-93f6-5ca0803c9e78 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.296271] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb63f30-111b-4ff3-a689-8a68bb9bc0e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.302548] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4589420-0c68-4c58-9f5d-d34733e30be3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.309050] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04fbb14a-432c-45f8-bc41-c84224202f31 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.321846] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f6618f-76b8-4d31-afc6-7f1fdc6c609a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.327537] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f053c9-c874-4791-92ac-1a896cd2c0db {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.358035] env[61243]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-c518ba67-98fa-41b4-bce5-4cdfb1be082a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.362852] env[61243]: DEBUG nova.virt.vmwareapi.driver [None req-f884f6a0-eb96-431b-9d32-d08740e4a07a None None] Extension org.openstack.compute already exists. {{(pid=61243) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 472.365454] env[61243]: INFO nova.compute.provider_config [None req-f884f6a0-eb96-431b-9d32-d08740e4a07a None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 472.869123] env[61243]: DEBUG nova.context [None req-f884f6a0-eb96-431b-9d32-d08740e4a07a None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),0bcfb6a2-0168-4a88-88b1-4c9a9b7dc3b7(cell1) {{(pid=61243) load_cells /opt/stack/nova/nova/context.py:464}} [ 472.871405] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 472.871702] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 472.872466] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 472.872920] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] Acquiring lock "0bcfb6a2-0168-4a88-88b1-4c9a9b7dc3b7" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 472.873128] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] Lock "0bcfb6a2-0168-4a88-88b1-4c9a9b7dc3b7" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 472.874133] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] Lock "0bcfb6a2-0168-4a88-88b1-4c9a9b7dc3b7" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 472.895758] env[61243]: INFO dbcounter [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] Registered counter for database nova_cell0 [ 472.904286] env[61243]: INFO dbcounter [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] Registered counter for database nova_cell1 [ 472.907497] env[61243]: DEBUG oslo_db.sqlalchemy.engines [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61243) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 472.907858] env[61243]: DEBUG oslo_db.sqlalchemy.engines [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61243) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 472.912917] env[61243]: ERROR nova.db.main.api [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 472.912917] env[61243]: result = function(*args, **kwargs) [ 472.912917] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 472.912917] env[61243]: return func(*args, **kwargs) [ 472.912917] env[61243]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 472.912917] env[61243]: result = fn(*args, **kwargs) [ 472.912917] env[61243]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 472.912917] env[61243]: return f(*args, **kwargs) [ 472.912917] env[61243]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 472.912917] env[61243]: return db.service_get_minimum_version(context, binaries) [ 472.912917] env[61243]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 472.912917] env[61243]: _check_db_access() [ 472.912917] env[61243]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 472.912917] env[61243]: stacktrace = ''.join(traceback.format_stack()) [ 472.912917] env[61243]: [ 472.913753] env[61243]: ERROR nova.db.main.api [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 472.913753] env[61243]: result = function(*args, **kwargs) [ 472.913753] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 472.913753] env[61243]: return func(*args, **kwargs) [ 472.913753] env[61243]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 472.913753] env[61243]: result = fn(*args, **kwargs) [ 472.913753] env[61243]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 472.913753] env[61243]: return f(*args, **kwargs) [ 472.913753] env[61243]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 472.913753] env[61243]: return db.service_get_minimum_version(context, binaries) [ 472.913753] env[61243]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 472.913753] env[61243]: _check_db_access() [ 472.913753] env[61243]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 472.913753] env[61243]: stacktrace = ''.join(traceback.format_stack()) [ 472.913753] env[61243]: [ 472.914161] env[61243]: WARNING nova.objects.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 472.914301] env[61243]: WARNING nova.objects.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] Failed to get minimum service version for cell 0bcfb6a2-0168-4a88-88b1-4c9a9b7dc3b7 [ 472.914724] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] Acquiring lock "singleton_lock" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 472.914888] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] Acquired lock "singleton_lock" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 472.915153] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] Releasing lock "singleton_lock" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 472.915478] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] Full set of CONF: {{(pid=61243) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 472.915623] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ******************************************************************************** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 472.915754] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] Configuration options gathered from: {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 472.915891] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 472.916100] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 472.916233] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ================================================================================ {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 472.916441] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] allow_resize_to_same_host = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.916613] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] arq_binding_timeout = 300 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.916749] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] backdoor_port = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.916940] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] backdoor_socket = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.917056] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] block_device_allocate_retries = 60 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.917226] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] block_device_allocate_retries_interval = 3 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.917398] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cert = self.pem {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.917570] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.917741] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] compute_monitors = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.917913] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] config_dir = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.918099] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] config_drive_format = iso9660 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.918237] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.918407] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] config_source = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.918579] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] console_host = devstack {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.918745] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] control_exchange = nova {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.918909] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cpu_allocation_ratio = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.919082] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] daemon = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.919257] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] debug = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.919417] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] default_access_ip_network_name = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.919582] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] default_availability_zone = nova {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.919767] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] default_ephemeral_format = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.919942] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] default_green_pool_size = 1000 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.920194] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.920364] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] default_schedule_zone = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.920526] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] disk_allocation_ratio = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.920691] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] enable_new_services = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.920873] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] enabled_apis = ['osapi_compute'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.921054] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] enabled_ssl_apis = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.921224] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] flat_injected = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.921387] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] force_config_drive = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.921549] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] force_raw_images = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.921722] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] graceful_shutdown_timeout = 5 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.921887] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] heal_instance_info_cache_interval = 60 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.922147] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] host = cpu-1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.922339] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.922583] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.922779] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.923015] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.923190] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] instance_build_timeout = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.923359] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] instance_delete_interval = 300 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.923531] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] instance_format = [instance: %(uuid)s] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.923702] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] instance_name_template = instance-%08x {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.923868] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] instance_usage_audit = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.924056] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] instance_usage_audit_period = month {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.924232] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.924403] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.924571] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] internal_service_availability_zone = internal {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.924731] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] key = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.924896] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] live_migration_retry_count = 30 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.925076] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] log_color = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.925246] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] log_config_append = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.925416] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.925578] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] log_dir = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.925741] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] log_file = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.925872] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] log_options = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.926046] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] log_rotate_interval = 1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.926229] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] log_rotate_interval_type = days {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.926391] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] log_rotation_type = none {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.926523] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.926653] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.926820] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.926995] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.927131] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.927303] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] long_rpc_timeout = 1800 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.927465] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] max_concurrent_builds = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.927626] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] max_concurrent_live_migrations = 1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.927788] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] max_concurrent_snapshots = 5 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.927952] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] max_local_block_devices = 3 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.928166] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] max_logfile_count = 30 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.928287] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] max_logfile_size_mb = 200 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.928452] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] maximum_instance_delete_attempts = 5 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.928625] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] metadata_listen = 0.0.0.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.928794] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] metadata_listen_port = 8775 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.928966] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] metadata_workers = 2 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.929142] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] migrate_max_retries = -1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.929312] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] mkisofs_cmd = genisoimage {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.929522] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.929657] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] my_ip = 10.180.1.21 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.929856] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] network_allocate_retries = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.930057] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.930234] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.930403] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] osapi_compute_listen_port = 8774 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.930604] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] osapi_compute_unique_server_name_scope = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.930782] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] osapi_compute_workers = 2 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.930951] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] password_length = 12 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.931130] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] periodic_enable = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.931296] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] periodic_fuzzy_delay = 60 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.931465] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] pointer_model = usbtablet {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.931634] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] preallocate_images = none {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.931799] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] publish_errors = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.931934] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] pybasedir = /opt/stack/nova {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.932117] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ram_allocation_ratio = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.932287] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] rate_limit_burst = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.932458] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] rate_limit_except_level = CRITICAL {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.932696] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] rate_limit_interval = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.932889] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] reboot_timeout = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.933068] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] reclaim_instance_interval = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.933234] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] record = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.933407] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] reimage_timeout_per_gb = 60 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.933580] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] report_interval = 120 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.933748] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] rescue_timeout = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.933914] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] reserved_host_cpus = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.934091] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] reserved_host_disk_mb = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.934256] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] reserved_host_memory_mb = 512 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.934420] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] reserved_huge_pages = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.934586] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] resize_confirm_window = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.934749] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] resize_fs_using_block_device = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.934911] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] resume_guests_state_on_host_boot = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.935092] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.935260] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] rpc_response_timeout = 60 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.935425] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] run_external_periodic_tasks = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.935597] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] running_deleted_instance_action = reap {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.935762] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.935927] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] running_deleted_instance_timeout = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.936099] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] scheduler_instance_sync_interval = 120 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.936274] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] service_down_time = 720 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.936444] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] servicegroup_driver = db {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.936602] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] shell_completion = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.936763] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] shelved_offload_time = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.936926] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] shelved_poll_interval = 3600 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.937108] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] shutdown_timeout = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.937274] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] source_is_ipv6 = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.937436] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ssl_only = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.937690] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.937860] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] sync_power_state_interval = 600 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.938037] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] sync_power_state_pool_size = 1000 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.938219] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] syslog_log_facility = LOG_USER {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.938375] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] tempdir = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.938536] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] timeout_nbd = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.938707] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] transport_url = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.938871] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] update_resources_interval = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.939045] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] use_cow_images = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.939211] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] use_eventlog = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.939371] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] use_journal = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.939532] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] use_json = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.939709] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] use_rootwrap_daemon = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.939898] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] use_stderr = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.940078] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] use_syslog = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.940242] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vcpu_pin_set = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.940415] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vif_plugging_is_fatal = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.940587] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vif_plugging_timeout = 300 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.940789] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] virt_mkfs = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.940970] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] volume_usage_poll_interval = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.941150] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] watch_log_file = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.941327] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] web = /usr/share/spice-html5 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 472.941512] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.941683] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.941853] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.942039] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_concurrency.disable_process_locking = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.942356] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.942544] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.942792] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.942998] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.943197] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.943371] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.943558] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.auth_strategy = keystone {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.943750] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.compute_link_prefix = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.943973] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.944177] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.dhcp_domain = novalocal {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.944366] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.enable_instance_password = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.944557] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.glance_link_prefix = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.944731] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.944909] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.945087] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.instance_list_per_project_cells = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.945282] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.list_records_by_skipping_down_cells = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.945491] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.local_metadata_per_cell = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.945680] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.max_limit = 1000 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.945864] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.metadata_cache_expiration = 15 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.946057] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.neutron_default_tenant_id = default {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.946242] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.response_validation = warn {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.946417] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.use_neutron_default_nets = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.946590] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.946758] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.946932] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.947121] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.947385] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.vendordata_dynamic_targets = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.947467] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.vendordata_jsonfile_path = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.947646] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.947875] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.backend = dogpile.cache.memcached {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.948070] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.backend_argument = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.948254] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.config_prefix = cache.oslo {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.948450] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.dead_timeout = 60.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.948589] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.debug_cache_backend = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.948753] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.enable_retry_client = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.948920] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.enable_socket_keepalive = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.949105] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.enabled = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.949275] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.enforce_fips_mode = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.949443] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.expiration_time = 600 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.949608] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.hashclient_retry_attempts = 2 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.949806] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.949980] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.memcache_dead_retry = 300 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.950152] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.memcache_password = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.950321] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.950486] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.950652] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.memcache_pool_maxsize = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.950833] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.951019] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.memcache_sasl_enabled = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.951212] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.951384] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.951544] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.memcache_username = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.951714] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.proxies = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.951883] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.redis_db = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.952061] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.redis_password = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.952239] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.952419] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.952591] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.redis_server = localhost:6379 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.952759] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.redis_socket_timeout = 1.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.953037] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.redis_username = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.953217] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.retry_attempts = 2 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.953387] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.retry_delay = 0.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.953552] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.socket_keepalive_count = 1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.953729] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.socket_keepalive_idle = 1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.953894] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.socket_keepalive_interval = 1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.954069] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.tls_allowed_ciphers = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.954235] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.tls_cafile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.954394] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.tls_certfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.954558] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.tls_enabled = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.954718] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cache.tls_keyfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.954896] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cinder.auth_section = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.955088] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cinder.auth_type = password {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.955256] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cinder.cafile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.955512] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.955697] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cinder.certfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.955870] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cinder.collect_timing = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.956050] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cinder.cross_az_attach = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.956223] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cinder.debug = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.956389] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cinder.endpoint_template = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.956557] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cinder.http_retries = 3 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.956729] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cinder.insecure = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.956893] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cinder.keyfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.957080] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cinder.os_region_name = RegionOne {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.957252] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cinder.split_loggers = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.957416] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cinder.timeout = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.957596] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.957781] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] compute.cpu_dedicated_set = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.957961] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] compute.cpu_shared_set = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.958146] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] compute.image_type_exclude_list = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.958316] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.958483] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.958650] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.958816] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.958990] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.959172] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] compute.resource_provider_association_refresh = 300 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.959337] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.959501] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] compute.shutdown_retry_interval = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.959689] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.959904] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] conductor.workers = 2 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.960108] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] console.allowed_origins = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.960276] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] console.ssl_ciphers = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.960454] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] console.ssl_minimum_version = default {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.960628] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] consoleauth.enforce_session_timeout = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.960835] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] consoleauth.token_ttl = 600 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.961034] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.cafile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.961204] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.certfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.961369] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.collect_timing = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.961533] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.connect_retries = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.961694] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.connect_retry_delay = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.961858] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.endpoint_override = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.962032] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.insecure = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.962197] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.keyfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.962359] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.max_version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.962518] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.min_version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.962676] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.region_name = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.962834] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.retriable_status_codes = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.963097] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.service_name = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.963288] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.service_type = accelerator {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.963456] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.split_loggers = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.963617] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.status_code_retries = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.963812] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.status_code_retry_delay = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.964007] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.timeout = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.964203] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.964373] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] cyborg.version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.964561] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.backend = sqlalchemy {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.964738] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.connection = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.964912] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.connection_debug = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.965098] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.connection_parameters = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.965268] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.connection_recycle_time = 3600 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.965436] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.connection_trace = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.965604] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.db_inc_retry_interval = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.965772] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.db_max_retries = 20 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.965940] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.db_max_retry_interval = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.966201] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.db_retry_interval = 1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.966406] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.max_overflow = 50 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.966580] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.max_pool_size = 5 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.966747] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.max_retries = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.966925] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.967101] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.mysql_wsrep_sync_wait = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.967268] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.pool_timeout = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.967436] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.retry_interval = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.967599] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.slave_connection = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.967787] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.sqlite_synchronous = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.967965] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] database.use_db_reconnect = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.968162] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.backend = sqlalchemy {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.968335] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.connection = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.968504] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.connection_debug = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.968672] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.connection_parameters = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.968836] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.connection_recycle_time = 3600 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.969009] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.connection_trace = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.969222] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.db_inc_retry_interval = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.969400] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.db_max_retries = 20 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.969569] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.db_max_retry_interval = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.969757] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.db_retry_interval = 1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.969937] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.max_overflow = 50 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.970119] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.max_pool_size = 5 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.970289] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.max_retries = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.970463] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.970628] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.970790] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.pool_timeout = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.970956] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.retry_interval = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.971132] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.slave_connection = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.971298] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] api_database.sqlite_synchronous = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.971479] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] devices.enabled_mdev_types = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.971658] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.971830] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.971999] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ephemeral_storage_encryption.enabled = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.972210] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.972396] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.api_servers = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.972567] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.cafile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.972733] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.certfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.972903] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.collect_timing = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.973152] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.connect_retries = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.973341] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.connect_retry_delay = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.973511] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.debug = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.973684] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.default_trusted_certificate_ids = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.973853] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.enable_certificate_validation = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.974028] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.enable_rbd_download = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.974198] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.endpoint_override = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.974370] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.insecure = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.974538] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.keyfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.974704] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.max_version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.974869] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.min_version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.975047] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.num_retries = 3 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.975259] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.rbd_ceph_conf = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.975436] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.rbd_connect_timeout = 5 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.975610] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.rbd_pool = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.975782] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.rbd_user = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.975947] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.region_name = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.976123] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.retriable_status_codes = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.976288] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.service_name = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.976460] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.service_type = image {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.976625] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.split_loggers = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.976793] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.status_code_retries = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.976954] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.status_code_retry_delay = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.977128] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.timeout = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.977314] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.977483] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.verify_glance_signatures = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.977648] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] glance.version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.977847] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] guestfs.debug = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.978025] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] mks.enabled = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.978414] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.978614] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] image_cache.manager_interval = 2400 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.978789] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] image_cache.precache_concurrency = 1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.978968] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] image_cache.remove_unused_base_images = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.979155] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.979329] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.979506] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] image_cache.subdirectory_name = _base {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.979713] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.api_max_retries = 60 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.979888] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.api_retry_interval = 2 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.980067] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.auth_section = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.980239] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.auth_type = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.980401] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.cafile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.980562] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.certfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.980758] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.collect_timing = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.980951] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.conductor_group = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.981131] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.connect_retries = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.981298] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.connect_retry_delay = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.981460] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.endpoint_override = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.981627] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.insecure = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.981788] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.keyfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.981951] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.max_version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.982127] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.min_version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.982295] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.peer_list = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.982456] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.region_name = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.982615] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.retriable_status_codes = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.982780] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.serial_console_state_timeout = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.982942] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.service_name = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.983193] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.service_type = baremetal {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.983385] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.shard = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.983559] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.split_loggers = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.983730] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.status_code_retries = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.983925] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.status_code_retry_delay = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.984106] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.timeout = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.984297] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.984463] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ironic.version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.984650] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.984829] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] key_manager.fixed_key = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.985025] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.985195] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.barbican_api_version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.985358] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.barbican_endpoint = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.985533] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.barbican_endpoint_type = public {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.985695] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.barbican_region_name = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.985862] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.cafile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.986031] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.certfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.986203] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.collect_timing = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.986368] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.insecure = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.986527] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.keyfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.986696] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.number_of_retries = 60 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.986899] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.retry_delay = 1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.987089] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.send_service_user_token = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.987270] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.split_loggers = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.987421] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.timeout = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.987585] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.verify_ssl = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.987764] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican.verify_ssl_path = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.987966] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican_service_user.auth_section = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.988122] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican_service_user.auth_type = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.988285] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican_service_user.cafile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.988443] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican_service_user.certfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.988609] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican_service_user.collect_timing = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.988770] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican_service_user.insecure = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.988931] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican_service_user.keyfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.989104] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican_service_user.split_loggers = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.989286] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] barbican_service_user.timeout = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.989468] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vault.approle_role_id = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.989630] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vault.approle_secret_id = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.989852] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vault.kv_mountpoint = secret {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.990043] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vault.kv_path = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.990220] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vault.kv_version = 2 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.990384] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vault.namespace = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.990546] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vault.root_token_id = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.990708] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vault.ssl_ca_crt_file = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.990878] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vault.timeout = 60.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.991052] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vault.use_ssl = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.991228] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.991407] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.auth_section = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.991572] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.auth_type = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.991732] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.cafile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.991894] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.certfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.992071] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.collect_timing = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.992233] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.connect_retries = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.992394] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.connect_retry_delay = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.992553] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.endpoint_override = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.992715] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.insecure = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.992912] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.keyfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.993092] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.max_version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.993324] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.min_version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.993507] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.region_name = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.993674] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.retriable_status_codes = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.993836] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.service_name = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.994020] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.service_type = identity {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.994191] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.split_loggers = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.994353] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.status_code_retries = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.994514] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.status_code_retry_delay = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.994673] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.timeout = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.995802] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.995802] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] keystone.version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.995802] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.connection_uri = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.995802] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.cpu_mode = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.995802] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.995802] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.cpu_models = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.996262] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.cpu_power_governor_high = performance {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.996262] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.996262] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.cpu_power_management = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.996429] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.996570] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.device_detach_attempts = 8 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.996741] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.device_detach_timeout = 20 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.996907] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.disk_cachemodes = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.998735] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.disk_prefix = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.998735] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.enabled_perf_events = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.998735] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.file_backed_memory = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.998735] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.gid_maps = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.998735] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.hw_disk_discard = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.998735] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.hw_machine_type = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.998735] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.images_rbd_ceph_conf = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.999156] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.999156] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.999156] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.images_rbd_glance_store_name = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.999156] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.images_rbd_pool = rbd {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.999156] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.images_type = default {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.999156] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.images_volume_group = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.999313] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.inject_key = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.999570] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.inject_partition = -2 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.999570] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.inject_password = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.999726] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.iscsi_iface = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 472.999933] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.iser_use_multipath = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.000121] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.000295] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.000465] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.live_migration_downtime = 500 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.000631] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.000820] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.000991] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.live_migration_inbound_addr = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.001171] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.001342] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.001508] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.live_migration_scheme = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.001686] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.live_migration_timeout_action = abort {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.001872] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.live_migration_tunnelled = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.002066] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.live_migration_uri = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.002239] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.live_migration_with_native_tls = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.002404] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.max_queues = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.002569] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.002813] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.002983] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.nfs_mount_options = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.003368] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.003577] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.003755] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.003926] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.004109] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.004281] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.num_pcie_ports = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.004453] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.004623] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.pmem_namespaces = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.004804] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.quobyte_client_cfg = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.005211] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.005415] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.005593] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.005765] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.005979] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.rbd_secret_uuid = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.006174] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.rbd_user = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.006347] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.006523] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.006688] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.rescue_image_id = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.006851] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.rescue_kernel_id = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.007015] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.rescue_ramdisk_id = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.007192] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.007354] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.rx_queue_size = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.007524] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.smbfs_mount_options = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.007808] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.008024] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.snapshot_compression = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.008199] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.snapshot_image_format = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.008421] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.008591] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.sparse_logical_volumes = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.008756] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.swtpm_enabled = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.008930] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.swtpm_group = tss {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.009112] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.swtpm_user = tss {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.009286] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.sysinfo_serial = unique {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.009445] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.tb_cache_size = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.009615] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.tx_queue_size = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.009798] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.uid_maps = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.009981] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.use_virtio_for_bridges = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.010173] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.virt_type = kvm {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.010348] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.volume_clear = zero {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.010513] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.volume_clear_size = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.010687] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.volume_use_multipath = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.010890] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.vzstorage_cache_path = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.011082] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.011257] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.011428] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.011598] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.011876] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.012090] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.vzstorage_mount_user = stack {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.012277] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.012460] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.auth_section = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.012639] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.auth_type = password {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.012804] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.cafile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.012968] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.certfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.013150] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.collect_timing = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.013315] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.connect_retries = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.013550] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.connect_retry_delay = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.013761] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.default_floating_pool = public {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.013949] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.endpoint_override = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.014133] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.extension_sync_interval = 600 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.014302] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.http_retries = 3 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.014465] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.insecure = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.014628] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.keyfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.014792] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.max_version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.014968] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.015146] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.min_version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.015318] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.ovs_bridge = br-int {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.015485] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.physnets = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.015656] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.region_name = RegionOne {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.015818] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.retriable_status_codes = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.015990] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.service_metadata_proxy = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.016163] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.service_name = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.016335] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.service_type = network {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.016498] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.split_loggers = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.016658] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.status_code_retries = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.016844] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.status_code_retry_delay = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.017036] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.timeout = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.017229] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.017396] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] neutron.version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.017576] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] notifications.bdms_in_notifications = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.017758] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] notifications.default_level = INFO {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.017935] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] notifications.notification_format = unversioned {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.018114] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] notifications.notify_on_state_change = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.018297] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.018478] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] pci.alias = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.018652] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] pci.device_spec = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.018818] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] pci.report_in_placement = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.018995] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.auth_section = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.019184] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.auth_type = password {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.019357] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.019521] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.cafile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.019779] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.certfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.019899] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.collect_timing = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.020089] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.connect_retries = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.020258] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.connect_retry_delay = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.020421] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.default_domain_id = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.020582] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.default_domain_name = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.020744] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.domain_id = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.020905] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.domain_name = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.021076] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.endpoint_override = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.021241] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.insecure = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.021402] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.keyfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.021563] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.max_version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.021721] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.min_version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.021895] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.password = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.022087] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.project_domain_id = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.022278] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.project_domain_name = Default {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.022456] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.project_id = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.022632] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.project_name = service {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.022836] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.region_name = RegionOne {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.023036] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.retriable_status_codes = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.023209] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.service_name = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.023384] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.service_type = placement {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.023619] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.split_loggers = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.023803] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.status_code_retries = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.023971] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.status_code_retry_delay = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.024151] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.system_scope = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.024319] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.timeout = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.024484] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.trust_id = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.024648] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.user_domain_id = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.024822] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.user_domain_name = Default {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.024984] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.user_id = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.025176] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.username = nova {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.025361] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.025526] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] placement.version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.025706] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] quota.cores = 20 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.025899] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] quota.count_usage_from_placement = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.026096] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.026279] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] quota.injected_file_content_bytes = 10240 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.026453] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] quota.injected_file_path_length = 255 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.026623] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] quota.injected_files = 5 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.026793] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] quota.instances = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.026963] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] quota.key_pairs = 100 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.027145] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] quota.metadata_items = 128 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.027316] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] quota.ram = 51200 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.027481] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] quota.recheck_quota = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.027650] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] quota.server_group_members = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.027816] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] quota.server_groups = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.027992] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.028171] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.028336] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] scheduler.image_metadata_prefilter = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.028503] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.028670] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] scheduler.max_attempts = 3 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.028862] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] scheduler.max_placement_results = 1000 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.029062] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.029236] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.029402] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.029577] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] scheduler.workers = 2 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.029806] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.030095] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.030187] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.030345] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.030514] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.030702] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.030879] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.031055] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.031230] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.host_subset_size = 1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.031397] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.031557] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.031721] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.031918] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.isolated_hosts = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.032131] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.isolated_images = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.032308] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.032476] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.032645] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.032814] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.pci_in_placement = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.032985] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.033169] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.033336] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.033502] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.033741] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.033930] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.034111] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.track_instance_changes = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.034296] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.034471] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] metrics.required = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.034637] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] metrics.weight_multiplier = 1.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.034809] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.035028] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] metrics.weight_setting = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.035362] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.035543] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] serial_console.enabled = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.035727] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] serial_console.port_range = 10000:20000 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.035904] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.036087] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.036261] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] serial_console.serialproxy_port = 6083 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.036431] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] service_user.auth_section = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.036610] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] service_user.auth_type = password {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.036773] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] service_user.cafile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.036936] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] service_user.certfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.037113] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] service_user.collect_timing = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.037280] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] service_user.insecure = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.037440] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] service_user.keyfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.037638] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] service_user.send_service_user_token = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.037894] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] service_user.split_loggers = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.038110] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] service_user.timeout = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.038294] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] spice.agent_enabled = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.038465] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] spice.enabled = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.038787] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.038980] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.039166] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] spice.html5proxy_port = 6082 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.039330] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] spice.image_compression = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.039491] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] spice.jpeg_compression = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.039651] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] spice.playback_compression = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.039847] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] spice.require_secure = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.040073] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] spice.server_listen = 127.0.0.1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.040204] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.040369] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] spice.streaming_mode = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.040528] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] spice.zlib_compression = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.040707] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] upgrade_levels.baseapi = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.040916] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] upgrade_levels.compute = auto {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.041099] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] upgrade_levels.conductor = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.041263] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] upgrade_levels.scheduler = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.041431] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.041597] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.041757] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vendordata_dynamic_auth.cafile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.041920] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vendordata_dynamic_auth.certfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.042124] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.042306] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vendordata_dynamic_auth.insecure = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.042471] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.042638] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.042801] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vendordata_dynamic_auth.timeout = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.042982] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.api_retry_count = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.043158] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.ca_file = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.043331] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.043500] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.cluster_name = testcl1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.043665] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.connection_pool_size = 10 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.043944] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.console_delay_seconds = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.044144] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.datastore_regex = ^datastore.* {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.044355] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.044530] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.host_password = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.044700] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.host_port = 443 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.044872] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.host_username = administrator@vsphere.local {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.045055] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.insecure = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.045222] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.integration_bridge = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.045389] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.maximum_objects = 100 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.045549] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.pbm_default_policy = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.045714] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.pbm_enabled = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.045876] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.pbm_wsdl_location = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.046056] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.046220] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.serial_port_proxy_uri = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.046378] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.serial_port_service_uri = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.046546] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.task_poll_interval = 0.5 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.046720] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.use_linked_clone = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.046929] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.vnc_keymap = en-us {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.047120] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.vnc_port = 5900 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.047290] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vmware.vnc_port_total = 10000 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.047478] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vnc.auth_schemes = ['none'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.047658] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vnc.enabled = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.047950] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.048147] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.048322] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vnc.novncproxy_port = 6080 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.048500] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vnc.server_listen = 127.0.0.1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.048673] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.048837] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vnc.vencrypt_ca_certs = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.048999] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vnc.vencrypt_client_cert = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.049174] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vnc.vencrypt_client_key = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.049354] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.049519] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.disable_deep_image_inspection = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.049691] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.049895] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.050080] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.050280] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.disable_rootwrap = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.050411] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.enable_numa_live_migration = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.050576] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.050747] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.050904] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.051076] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.libvirt_disable_apic = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.051241] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.051406] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.051569] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.051732] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.051894] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.052100] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.052287] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.052513] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.052616] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.052813] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.053032] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.053212] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] wsgi.client_socket_timeout = 900 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.053383] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] wsgi.default_pool_size = 1000 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.053553] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] wsgi.keep_alive = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.053722] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] wsgi.max_header_line = 16384 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.053965] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.054152] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] wsgi.ssl_ca_file = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.054317] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] wsgi.ssl_cert_file = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.054481] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] wsgi.ssl_key_file = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.054664] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] wsgi.tcp_keepidle = 600 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.054861] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.055042] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] zvm.ca_file = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.055210] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] zvm.cloud_connector_url = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.055497] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.055672] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] zvm.reachable_timeout = 300 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.055887] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_policy.enforce_new_defaults = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.056293] env[61243]: WARNING oslo_config.cfg [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 473.056498] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_policy.enforce_scope = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.056681] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_policy.policy_default_rule = default {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.056868] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.057063] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_policy.policy_file = policy.yaml {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.057242] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.057405] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.057570] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.057729] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.057893] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.058075] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.058257] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.058435] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] profiler.connection_string = messaging:// {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.058606] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] profiler.enabled = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.058776] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] profiler.es_doc_type = notification {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.058984] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] profiler.es_scroll_size = 10000 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.059178] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] profiler.es_scroll_time = 2m {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.059346] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] profiler.filter_error_trace = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.059519] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] profiler.hmac_keys = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.059691] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] profiler.sentinel_service_name = mymaster {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.059889] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] profiler.socket_timeout = 0.1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.060126] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] profiler.trace_requests = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.060305] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] profiler.trace_sqlalchemy = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.060490] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] profiler_jaeger.process_tags = {} {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.060655] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] profiler_jaeger.service_name_prefix = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.060849] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] profiler_otlp.service_name_prefix = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.061033] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] remote_debug.host = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.061201] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] remote_debug.port = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.061384] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.061550] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.061718] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.061906] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.062115] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.062294] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.062461] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.062627] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.062791] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.062964] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.063137] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.063311] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.063481] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.063651] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.063822] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.064083] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.064263] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.064443] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.064609] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.064775] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.064991] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.065185] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.065353] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.065522] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.065689] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.065854] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.066027] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.066195] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.066368] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.066536] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.ssl = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.066710] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.066884] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.067058] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.067269] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.067512] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.067694] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.067889] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.068072] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_notifications.retry = -1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.068264] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.068444] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.068627] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.auth_section = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.068793] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.auth_type = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.068955] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.cafile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.069132] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.certfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.069299] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.collect_timing = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.069460] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.connect_retries = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.069620] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.connect_retry_delay = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.069806] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.endpoint_id = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.069977] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.endpoint_override = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.070159] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.insecure = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.070416] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.keyfile = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.070540] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.max_version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.070703] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.min_version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.070885] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.region_name = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.071065] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.retriable_status_codes = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.071232] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.service_name = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.071395] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.service_type = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.071560] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.split_loggers = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.071721] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.status_code_retries = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.071884] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.status_code_retry_delay = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.072055] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.timeout = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.072225] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.valid_interfaces = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.072425] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_limit.version = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.072602] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_reports.file_event_handler = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.072769] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.072932] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] oslo_reports.log_dir = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.073121] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.073306] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.073486] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.073658] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.073828] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.074062] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.074264] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.074431] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vif_plug_ovs_privileged.group = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.074593] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.074764] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.074934] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.075104] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] vif_plug_ovs_privileged.user = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.075281] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.075466] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.075641] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.075817] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.075992] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.076175] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.076375] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.076561] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.076766] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.076962] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_vif_ovs.isolate_vif = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.077153] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.077326] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.077513] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.077667] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.077837] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] os_vif_ovs.per_port_bridge = False {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.078022] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] privsep_osbrick.capabilities = [21] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.078184] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] privsep_osbrick.group = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.078347] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] privsep_osbrick.helper_command = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.078514] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.078683] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.078845] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] privsep_osbrick.user = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.079032] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.079201] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] nova_sys_admin.group = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.079405] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] nova_sys_admin.helper_command = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.079586] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.079779] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.079954] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] nova_sys_admin.user = None {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 473.080099] env[61243]: DEBUG oslo_service.service [None req-8a58f721-fca6-47a3-b3d3-c26f0103def1 None None] ******************************************************************************** {{(pid=61243) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 473.080645] env[61243]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 473.584199] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Getting list of instances from cluster (obj){ [ 473.584199] env[61243]: value = "domain-c8" [ 473.584199] env[61243]: _type = "ClusterComputeResource" [ 473.584199] env[61243]: } {{(pid=61243) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 473.585368] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea06a6e-70e5-425b-b4bb-acf4868c0b9c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 473.594449] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Got total of 0 instances {{(pid=61243) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 473.594994] env[61243]: WARNING nova.virt.vmwareapi.driver [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 473.595501] env[61243]: INFO nova.virt.node [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Generated node identity 8f3c282f-58fe-4c5d-80db-5a142cf023e1 [ 473.595737] env[61243]: INFO nova.virt.node [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Wrote node identity 8f3c282f-58fe-4c5d-80db-5a142cf023e1 to /opt/stack/data/n-cpu-1/compute_id [ 474.102016] env[61243]: WARNING nova.compute.manager [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Compute nodes ['8f3c282f-58fe-4c5d-80db-5a142cf023e1'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 475.107892] env[61243]: INFO nova.compute.manager [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 476.113779] env[61243]: WARNING nova.compute.manager [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 476.114146] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 476.114298] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 476.114458] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 476.114633] env[61243]: DEBUG nova.compute.resource_tracker [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61243) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 476.115551] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fdd29a7-22ef-4106-a7c3-a835570e0e02 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.123840] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1edb34c-27a1-487b-84fe-112ff4f6d5fb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.137923] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde7c32d-0033-4a87-a474-5444407c069a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.144253] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757069fb-bcad-4c6a-98f6-c2620750f2d5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.171591] env[61243]: DEBUG nova.compute.resource_tracker [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181537MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=61243) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 476.171737] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 476.171931] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 476.674823] env[61243]: WARNING nova.compute.resource_tracker [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] No compute node record for cpu-1:8f3c282f-58fe-4c5d-80db-5a142cf023e1: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 8f3c282f-58fe-4c5d-80db-5a142cf023e1 could not be found. [ 477.179020] env[61243]: INFO nova.compute.resource_tracker [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 [ 478.687690] env[61243]: DEBUG nova.compute.resource_tracker [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 478.688063] env[61243]: DEBUG nova.compute.resource_tracker [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 478.835762] env[61243]: INFO nova.scheduler.client.report [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] [req-fc721f9f-a73c-46c3-a936-4086d310a1d1] Created resource provider record via placement API for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 478.851801] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e635b819-b6e6-470b-b445-cd2349ac35e4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.859444] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b6951b-fb75-482b-a468-e79678bce836 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.888055] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b19f69-3e0d-46a2-86ce-b71cb2f1de56 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.894561] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e17481-8055-46b7-9e0a-8ab26a3f12c7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.906811] env[61243]: DEBUG nova.compute.provider_tree [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 479.443264] env[61243]: DEBUG nova.scheduler.client.report [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 479.443494] env[61243]: DEBUG nova.compute.provider_tree [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 0 to 1 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 479.443636] env[61243]: DEBUG nova.compute.provider_tree [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 479.492319] env[61243]: DEBUG nova.compute.provider_tree [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 1 to 2 during operation: update_traits {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 479.997460] env[61243]: DEBUG nova.compute.resource_tracker [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61243) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 479.997769] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.826s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 479.997886] env[61243]: DEBUG nova.service [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Creating RPC server for service compute {{(pid=61243) start /opt/stack/nova/nova/service.py:186}} [ 480.015448] env[61243]: DEBUG nova.service [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] Join ServiceGroup membership for this service compute {{(pid=61243) start /opt/stack/nova/nova/service.py:203}} [ 480.015644] env[61243]: DEBUG nova.servicegroup.drivers.db [None req-f55733d2-2fe5-4901-ad69-4507a009ddcf None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61243) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 512.018699] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._sync_power_states {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 512.522336] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Getting list of instances from cluster (obj){ [ 512.522336] env[61243]: value = "domain-c8" [ 512.522336] env[61243]: _type = "ClusterComputeResource" [ 512.522336] env[61243]: } {{(pid=61243) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 512.523965] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a17da6-cc60-4510-be48-43fabd097d89 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.534985] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Got total of 0 instances {{(pid=61243) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 512.535264] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 512.535591] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Getting list of instances from cluster (obj){ [ 512.535591] env[61243]: value = "domain-c8" [ 512.535591] env[61243]: _type = "ClusterComputeResource" [ 512.535591] env[61243]: } {{(pid=61243) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 512.538680] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292f12e6-cb88-4729-b814-06a8403790a1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.547106] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Got total of 0 instances {{(pid=61243) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 517.889418] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Acquiring lock "35aeb5b2-ca5f-457b-9545-05ae609e5ac9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.889709] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Lock "35aeb5b2-ca5f-457b-9545-05ae609e5ac9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.393296] env[61243]: DEBUG nova.compute.manager [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 518.869967] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Acquiring lock "906184f9-27b9-4ce5-aa35-d2a4cdaecae3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.870424] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Lock "906184f9-27b9-4ce5-aa35-d2a4cdaecae3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.940238] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.940238] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.942425] env[61243]: INFO nova.compute.claims [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 519.228391] env[61243]: DEBUG oslo_concurrency.lockutils [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquiring lock "1b1b2bac-67af-4921-a49a-429017ef7a3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.228619] env[61243]: DEBUG oslo_concurrency.lockutils [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Lock "1b1b2bac-67af-4921-a49a-429017ef7a3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.375285] env[61243]: DEBUG nova.compute.manager [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 519.678603] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Acquiring lock "ffae890f-e876-4382-a21e-35b7e9c05da9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.678603] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Lock "ffae890f-e876-4382-a21e-35b7e9c05da9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.734794] env[61243]: DEBUG nova.compute.manager [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 519.750099] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Acquiring lock "6c4e205a-6f9d-4439-add7-376fe718d189" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.753905] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Lock "6c4e205a-6f9d-4439-add7-376fe718d189" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.920168] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.063428] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7eb929a-f59f-4bd3-837f-389386263e53 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.075626] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b87bfc7-d40e-48a3-9910-2bbdb587c0a3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.108952] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d824ba5-3709-4631-b68b-24604aebfbf4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.118239] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c0ec19e-faed-4876-92a8-a5965fb953dd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.132223] env[61243]: DEBUG nova.compute.provider_tree [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 520.181512] env[61243]: DEBUG nova.compute.manager [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 520.257257] env[61243]: DEBUG nova.compute.manager [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 520.264947] env[61243]: DEBUG oslo_concurrency.lockutils [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.636900] env[61243]: DEBUG nova.scheduler.client.report [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 520.715582] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.798212] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.144154] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.204s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 521.144804] env[61243]: DEBUG nova.compute.manager [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 521.147527] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.227s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.149375] env[61243]: INFO nova.compute.claims [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 521.371471] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Acquiring lock "da977758-95d5-4fb6-aaa3-7b832404b527" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.371754] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Lock "da977758-95d5-4fb6-aaa3-7b832404b527" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.656178] env[61243]: DEBUG nova.compute.utils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 521.657079] env[61243]: DEBUG nova.compute.manager [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 521.657779] env[61243]: DEBUG nova.network.neutron [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 521.880447] env[61243]: DEBUG nova.compute.manager [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 522.057999] env[61243]: DEBUG nova.policy [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bdb88f6e2cb947b3b62b33aad6ebffee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '872c70a460b04e8e93112dc85a78e522', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 522.162757] env[61243]: DEBUG nova.compute.manager [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 522.293019] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8d7003-c4fc-4428-a344-9af5d14b6ce7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.301890] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea66a8b-b52a-46fb-8bd2-fc7998163d0c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.333500] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c695f3-9c3d-4f9c-a825-44d72881afed {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.341105] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d965c4a4-5c29-4a0a-89d9-c363671f0ebc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.356240] env[61243]: DEBUG nova.compute.provider_tree [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 522.405413] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.576250] env[61243]: DEBUG nova.network.neutron [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Successfully created port: 9171227f-a895-4cad-ab56-8f231377bcad {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 522.860765] env[61243]: DEBUG nova.scheduler.client.report [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 523.174135] env[61243]: DEBUG nova.compute.manager [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 523.206325] env[61243]: DEBUG nova.virt.hardware [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 523.206571] env[61243]: DEBUG nova.virt.hardware [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 523.206722] env[61243]: DEBUG nova.virt.hardware [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 523.206898] env[61243]: DEBUG nova.virt.hardware [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 523.207300] env[61243]: DEBUG nova.virt.hardware [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 523.207300] env[61243]: DEBUG nova.virt.hardware [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 523.207456] env[61243]: DEBUG nova.virt.hardware [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 523.207601] env[61243]: DEBUG nova.virt.hardware [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 523.207993] env[61243]: DEBUG nova.virt.hardware [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 523.208179] env[61243]: DEBUG nova.virt.hardware [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 523.208349] env[61243]: DEBUG nova.virt.hardware [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 523.209261] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab1b4bbf-a151-4a3e-a8ed-65ef7f63ba55 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.223029] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ecb81c5-5705-4f6c-b4d9-1e61d5cf4da0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.242047] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c163148-d655-4c3a-aa46-819513c50ff1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.367673] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.220s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 523.369152] env[61243]: DEBUG nova.compute.manager [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 523.374624] env[61243]: DEBUG oslo_concurrency.lockutils [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.111s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.376074] env[61243]: INFO nova.compute.claims [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 523.506924] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Acquiring lock "f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.507165] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Lock "f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.876048] env[61243]: DEBUG nova.compute.utils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 523.877590] env[61243]: DEBUG nova.compute.manager [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 523.877743] env[61243]: DEBUG nova.network.neutron [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 524.009700] env[61243]: DEBUG nova.compute.manager [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 524.078895] env[61243]: DEBUG nova.policy [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1dff0e19c58a44e6a2c23bcc0634f10a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '896c4c84e700417d81e2897540f86450', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 524.387024] env[61243]: DEBUG nova.compute.manager [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 524.537307] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.576263] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-540f9fc3-a6b8-4252-bddc-717afea32ff9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.586292] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ba212c1-a111-4a01-83f0-d1a56396d419 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.629779] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1eebc3-cc59-4d56-abab-f7e58e30f4f6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.638779] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef0d74d-d978-498e-ba83-53928c9de41a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.655906] env[61243]: DEBUG nova.compute.provider_tree [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 524.695040] env[61243]: ERROR nova.compute.manager [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9171227f-a895-4cad-ab56-8f231377bcad, please check neutron logs for more information. [ 524.695040] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 524.695040] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.695040] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 524.695040] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 524.695040] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 524.695040] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 524.695040] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 524.695040] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.695040] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 524.695040] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.695040] env[61243]: ERROR nova.compute.manager raise self.value [ 524.695040] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 524.695040] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 524.695040] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.695040] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 524.695531] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.695531] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 524.695531] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9171227f-a895-4cad-ab56-8f231377bcad, please check neutron logs for more information. [ 524.695531] env[61243]: ERROR nova.compute.manager [ 524.695531] env[61243]: Traceback (most recent call last): [ 524.695531] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 524.695531] env[61243]: listener.cb(fileno) [ 524.695531] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.695531] env[61243]: result = function(*args, **kwargs) [ 524.695531] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 524.695531] env[61243]: return func(*args, **kwargs) [ 524.695531] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.695531] env[61243]: raise e [ 524.695531] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.695531] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 524.695531] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 524.695531] env[61243]: created_port_ids = self._update_ports_for_instance( [ 524.695531] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 524.695531] env[61243]: with excutils.save_and_reraise_exception(): [ 524.695531] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.695531] env[61243]: self.force_reraise() [ 524.695531] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.695531] env[61243]: raise self.value [ 524.695531] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 524.695531] env[61243]: updated_port = self._update_port( [ 524.695531] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.695531] env[61243]: _ensure_no_port_binding_failure(port) [ 524.695531] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.695531] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 524.696676] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 9171227f-a895-4cad-ab56-8f231377bcad, please check neutron logs for more information. [ 524.696676] env[61243]: Removing descriptor: 15 [ 524.696676] env[61243]: ERROR nova.compute.manager [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9171227f-a895-4cad-ab56-8f231377bcad, please check neutron logs for more information. [ 524.696676] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Traceback (most recent call last): [ 524.696676] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 524.696676] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] yield resources [ 524.696676] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 524.696676] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] self.driver.spawn(context, instance, image_meta, [ 524.696676] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 524.696676] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 524.696676] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 524.696676] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] vm_ref = self.build_virtual_machine(instance, [ 524.697069] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 524.697069] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] vif_infos = vmwarevif.get_vif_info(self._session, [ 524.697069] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 524.697069] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] for vif in network_info: [ 524.697069] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 524.697069] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] return self._sync_wrapper(fn, *args, **kwargs) [ 524.697069] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 524.697069] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] self.wait() [ 524.697069] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 524.697069] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] self[:] = self._gt.wait() [ 524.697069] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 524.697069] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] return self._exit_event.wait() [ 524.697069] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 524.697400] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] result = hub.switch() [ 524.697400] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 524.697400] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] return self.greenlet.switch() [ 524.697400] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.697400] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] result = function(*args, **kwargs) [ 524.697400] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 524.697400] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] return func(*args, **kwargs) [ 524.697400] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.697400] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] raise e [ 524.697400] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.697400] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] nwinfo = self.network_api.allocate_for_instance( [ 524.697400] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 524.697400] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] created_port_ids = self._update_ports_for_instance( [ 524.698441] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 524.698441] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] with excutils.save_and_reraise_exception(): [ 524.698441] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.698441] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] self.force_reraise() [ 524.698441] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.698441] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] raise self.value [ 524.698441] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 524.698441] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] updated_port = self._update_port( [ 524.698441] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.698441] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] _ensure_no_port_binding_failure(port) [ 524.698441] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.698441] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] raise exception.PortBindingFailed(port_id=port['id']) [ 524.698896] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] nova.exception.PortBindingFailed: Binding failed for port 9171227f-a895-4cad-ab56-8f231377bcad, please check neutron logs for more information. [ 524.698896] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] [ 524.698896] env[61243]: INFO nova.compute.manager [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Terminating instance [ 524.701145] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Acquiring lock "refresh_cache-35aeb5b2-ca5f-457b-9545-05ae609e5ac9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.701145] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Acquired lock "refresh_cache-35aeb5b2-ca5f-457b-9545-05ae609e5ac9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.701145] env[61243]: DEBUG nova.network.neutron [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 525.162566] env[61243]: DEBUG nova.scheduler.client.report [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 525.246351] env[61243]: DEBUG nova.network.neutron [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Successfully created port: 6f27b665-bff9-433b-8b38-7954acda35e2 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 525.248756] env[61243]: DEBUG nova.network.neutron [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 525.395064] env[61243]: DEBUG nova.network.neutron [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.404797] env[61243]: DEBUG nova.compute.manager [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 525.537498] env[61243]: DEBUG nova.virt.hardware [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 525.537912] env[61243]: DEBUG nova.virt.hardware [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 525.537912] env[61243]: DEBUG nova.virt.hardware [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 525.538121] env[61243]: DEBUG nova.virt.hardware [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 525.538226] env[61243]: DEBUG nova.virt.hardware [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 525.538397] env[61243]: DEBUG nova.virt.hardware [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 525.538601] env[61243]: DEBUG nova.virt.hardware [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 525.538729] env[61243]: DEBUG nova.virt.hardware [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 525.539535] env[61243]: DEBUG nova.virt.hardware [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 525.539535] env[61243]: DEBUG nova.virt.hardware [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 525.539535] env[61243]: DEBUG nova.virt.hardware [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 525.540102] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81a21c02-cb91-4a1b-bbf5-3a97becd521b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.549509] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc77e690-3acd-4fa6-92e0-b18c03fc9d9d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.583316] env[61243]: DEBUG nova.compute.manager [req-4daed172-e8ae-400d-b363-bb922ae79650 req-6b16c1f6-2638-4113-80e9-a9b4e7091cb7 service nova] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Received event network-changed-9171227f-a895-4cad-ab56-8f231377bcad {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 525.583524] env[61243]: DEBUG nova.compute.manager [req-4daed172-e8ae-400d-b363-bb922ae79650 req-6b16c1f6-2638-4113-80e9-a9b4e7091cb7 service nova] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Refreshing instance network info cache due to event network-changed-9171227f-a895-4cad-ab56-8f231377bcad. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 525.583721] env[61243]: DEBUG oslo_concurrency.lockutils [req-4daed172-e8ae-400d-b363-bb922ae79650 req-6b16c1f6-2638-4113-80e9-a9b4e7091cb7 service nova] Acquiring lock "refresh_cache-35aeb5b2-ca5f-457b-9545-05ae609e5ac9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.673259] env[61243]: DEBUG oslo_concurrency.lockutils [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.298s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.673887] env[61243]: DEBUG nova.compute.manager [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 525.677655] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.963s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.680232] env[61243]: INFO nova.compute.claims [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 525.900683] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Releasing lock "refresh_cache-35aeb5b2-ca5f-457b-9545-05ae609e5ac9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.901155] env[61243]: DEBUG nova.compute.manager [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 525.905016] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 525.905016] env[61243]: DEBUG oslo_concurrency.lockutils [req-4daed172-e8ae-400d-b363-bb922ae79650 req-6b16c1f6-2638-4113-80e9-a9b4e7091cb7 service nova] Acquired lock "refresh_cache-35aeb5b2-ca5f-457b-9545-05ae609e5ac9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.905016] env[61243]: DEBUG nova.network.neutron [req-4daed172-e8ae-400d-b363-bb922ae79650 req-6b16c1f6-2638-4113-80e9-a9b4e7091cb7 service nova] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Refreshing network info cache for port 9171227f-a895-4cad-ab56-8f231377bcad {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 525.905016] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52a3a1aa-263e-46ec-908a-d0c10fa2d179 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.912450] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8344870-c357-45e8-a9c6-2b397552171a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.937266] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 35aeb5b2-ca5f-457b-9545-05ae609e5ac9 could not be found. [ 525.937716] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 525.938186] env[61243]: INFO nova.compute.manager [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 525.940082] env[61243]: DEBUG oslo.service.loopingcall [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 525.940082] env[61243]: DEBUG nova.compute.manager [-] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 525.940082] env[61243]: DEBUG nova.network.neutron [-] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 525.993931] env[61243]: DEBUG nova.network.neutron [-] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 526.186343] env[61243]: DEBUG nova.compute.utils [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 526.190012] env[61243]: DEBUG nova.compute.manager [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Not allocating networking since 'none' was specified. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 526.445483] env[61243]: DEBUG nova.network.neutron [req-4daed172-e8ae-400d-b363-bb922ae79650 req-6b16c1f6-2638-4113-80e9-a9b4e7091cb7 service nova] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 526.500668] env[61243]: DEBUG nova.network.neutron [-] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.693549] env[61243]: DEBUG nova.compute.manager [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 526.715079] env[61243]: DEBUG nova.network.neutron [req-4daed172-e8ae-400d-b363-bb922ae79650 req-6b16c1f6-2638-4113-80e9-a9b4e7091cb7 service nova] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.855017] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32fa394b-0a0b-4926-89bb-bb182dd5bd2e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.864163] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01535747-d2d8-443e-b010-1f4299e31dc9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.895030] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f254bc10-aea0-47cf-b4d6-d240249000ae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.903596] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-432d0f38-364d-47af-8a9b-6532651d7774 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.917085] env[61243]: DEBUG nova.compute.provider_tree [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 527.003664] env[61243]: INFO nova.compute.manager [-] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Took 1.06 seconds to deallocate network for instance. [ 527.010482] env[61243]: DEBUG nova.compute.claims [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 527.010482] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.218806] env[61243]: DEBUG oslo_concurrency.lockutils [req-4daed172-e8ae-400d-b363-bb922ae79650 req-6b16c1f6-2638-4113-80e9-a9b4e7091cb7 service nova] Releasing lock "refresh_cache-35aeb5b2-ca5f-457b-9545-05ae609e5ac9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.422496] env[61243]: DEBUG nova.scheduler.client.report [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 527.521368] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Acquiring lock "25a896cb-fd91-4f16-a8cd-6de58fcfcc27" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.523050] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Lock "25a896cb-fd91-4f16-a8cd-6de58fcfcc27" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.618608] env[61243]: ERROR nova.compute.manager [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6f27b665-bff9-433b-8b38-7954acda35e2, please check neutron logs for more information. [ 527.618608] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 527.618608] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.618608] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 527.618608] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 527.618608] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 527.618608] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 527.618608] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 527.618608] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.618608] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 527.618608] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.618608] env[61243]: ERROR nova.compute.manager raise self.value [ 527.618608] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 527.618608] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 527.618608] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.618608] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 527.619330] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.619330] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 527.619330] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6f27b665-bff9-433b-8b38-7954acda35e2, please check neutron logs for more information. [ 527.619330] env[61243]: ERROR nova.compute.manager [ 527.619330] env[61243]: Traceback (most recent call last): [ 527.619330] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 527.619330] env[61243]: listener.cb(fileno) [ 527.619330] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.619330] env[61243]: result = function(*args, **kwargs) [ 527.619330] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 527.619330] env[61243]: return func(*args, **kwargs) [ 527.619330] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 527.619330] env[61243]: raise e [ 527.619330] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.619330] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 527.619330] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 527.619330] env[61243]: created_port_ids = self._update_ports_for_instance( [ 527.619330] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 527.619330] env[61243]: with excutils.save_and_reraise_exception(): [ 527.619330] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.619330] env[61243]: self.force_reraise() [ 527.619330] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.619330] env[61243]: raise self.value [ 527.619330] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 527.619330] env[61243]: updated_port = self._update_port( [ 527.619330] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.619330] env[61243]: _ensure_no_port_binding_failure(port) [ 527.619330] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.619330] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 527.620390] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 6f27b665-bff9-433b-8b38-7954acda35e2, please check neutron logs for more information. [ 527.620390] env[61243]: Removing descriptor: 16 [ 527.620390] env[61243]: ERROR nova.compute.manager [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6f27b665-bff9-433b-8b38-7954acda35e2, please check neutron logs for more information. [ 527.620390] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Traceback (most recent call last): [ 527.620390] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 527.620390] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] yield resources [ 527.620390] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 527.620390] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] self.driver.spawn(context, instance, image_meta, [ 527.620390] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 527.620390] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 527.620390] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 527.620390] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] vm_ref = self.build_virtual_machine(instance, [ 527.622312] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 527.622312] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] vif_infos = vmwarevif.get_vif_info(self._session, [ 527.622312] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 527.622312] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] for vif in network_info: [ 527.622312] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 527.622312] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] return self._sync_wrapper(fn, *args, **kwargs) [ 527.622312] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 527.622312] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] self.wait() [ 527.622312] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 527.622312] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] self[:] = self._gt.wait() [ 527.622312] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 527.622312] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] return self._exit_event.wait() [ 527.622312] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 527.622688] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] result = hub.switch() [ 527.622688] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 527.622688] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] return self.greenlet.switch() [ 527.622688] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.622688] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] result = function(*args, **kwargs) [ 527.622688] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 527.622688] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] return func(*args, **kwargs) [ 527.622688] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 527.622688] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] raise e [ 527.622688] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.622688] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] nwinfo = self.network_api.allocate_for_instance( [ 527.622688] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 527.622688] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] created_port_ids = self._update_ports_for_instance( [ 527.623057] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 527.623057] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] with excutils.save_and_reraise_exception(): [ 527.623057] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.623057] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] self.force_reraise() [ 527.623057] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.623057] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] raise self.value [ 527.623057] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 527.623057] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] updated_port = self._update_port( [ 527.623057] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.623057] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] _ensure_no_port_binding_failure(port) [ 527.623057] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.623057] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] raise exception.PortBindingFailed(port_id=port['id']) [ 527.623338] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] nova.exception.PortBindingFailed: Binding failed for port 6f27b665-bff9-433b-8b38-7954acda35e2, please check neutron logs for more information. [ 527.623338] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] [ 527.623338] env[61243]: INFO nova.compute.manager [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Terminating instance [ 527.623338] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Acquiring lock "refresh_cache-906184f9-27b9-4ce5-aa35-d2a4cdaecae3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.623338] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Acquired lock "refresh_cache-906184f9-27b9-4ce5-aa35-d2a4cdaecae3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.623338] env[61243]: DEBUG nova.network.neutron [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 527.717958] env[61243]: DEBUG nova.compute.manager [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 527.756594] env[61243]: DEBUG nova.virt.hardware [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 527.756807] env[61243]: DEBUG nova.virt.hardware [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 527.756968] env[61243]: DEBUG nova.virt.hardware [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 527.757167] env[61243]: DEBUG nova.virt.hardware [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 527.757319] env[61243]: DEBUG nova.virt.hardware [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 527.757464] env[61243]: DEBUG nova.virt.hardware [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 527.757711] env[61243]: DEBUG nova.virt.hardware [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 527.757878] env[61243]: DEBUG nova.virt.hardware [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 527.758059] env[61243]: DEBUG nova.virt.hardware [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 527.759798] env[61243]: DEBUG nova.virt.hardware [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 527.760082] env[61243]: DEBUG nova.virt.hardware [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 527.760751] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c75138-2a16-4bcf-a44a-880276c1507f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.770713] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-688cb863-b1d8-4880-b399-ca761968623a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.798989] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Instance VIF info [] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 527.811812] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 527.815112] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e5e10eac-f13e-402b-b6b3-871bed1c64bd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.823942] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Created folder: OpenStack in parent group-v4. [ 527.825891] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Creating folder: Project (cf25c9d870264aeab7415e0c61535a3f). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 527.825891] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d48c462f-b1a0-4e43-9b08-afc353412971 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.837079] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Created folder: Project (cf25c9d870264aeab7415e0c61535a3f) in parent group-v285636. [ 527.839100] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Creating folder: Instances. Parent ref: group-v285637. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 527.839370] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3b5e8979-5421-470c-936a-d59d1b8c785d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.855135] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Created folder: Instances in parent group-v285637. [ 527.855135] env[61243]: DEBUG oslo.service.loopingcall [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 527.855135] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 527.855135] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0b7ef199-e7f4-4cb1-b443-bdd6c00b4ab2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.875024] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 527.875024] env[61243]: value = "task-1338684" [ 527.875024] env[61243]: _type = "Task" [ 527.875024] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.885503] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338684, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.931113] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.253s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.931802] env[61243]: DEBUG nova.compute.manager [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 527.934678] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.137s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.938034] env[61243]: INFO nova.compute.claims [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 528.032869] env[61243]: DEBUG nova.compute.manager [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 528.158582] env[61243]: DEBUG nova.network.neutron [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.273978] env[61243]: DEBUG nova.network.neutron [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.313632] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.314035] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.314284] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Starting heal instance info cache {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 528.314409] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Rebuilding the list of instances to heal {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 528.389874] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338684, 'name': CreateVM_Task, 'duration_secs': 0.294865} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.390117] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 528.391365] env[61243]: DEBUG oslo_vmware.service [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-859d737e-e25d-4cfb-aabf-7a560e519504 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.397826] env[61243]: DEBUG oslo_concurrency.lockutils [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.399167] env[61243]: DEBUG oslo_concurrency.lockutils [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.399167] env[61243]: DEBUG oslo_concurrency.lockutils [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 528.399167] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01a4cbef-6004-4bec-858e-817a9f10ee51 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.404305] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 528.404305] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527a9076-570b-ce1f-277b-d73482b66e95" [ 528.404305] env[61243]: _type = "Task" [ 528.404305] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.415707] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527a9076-570b-ce1f-277b-d73482b66e95, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.441348] env[61243]: DEBUG nova.compute.utils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 528.448012] env[61243]: DEBUG nova.compute.manager [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 528.448207] env[61243]: DEBUG nova.network.neutron [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 528.524759] env[61243]: DEBUG nova.policy [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3b49ae3d17b041868405785db34a58ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '57ea50dc7c7d4fd59f74ac433306e22e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 528.563233] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.777783] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Releasing lock "refresh_cache-906184f9-27b9-4ce5-aa35-d2a4cdaecae3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.778265] env[61243]: DEBUG nova.compute.manager [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 528.778478] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 528.778811] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f786e29-7d02-44e5-b2c9-ada0a1908b4a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.793229] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2caa4529-0c6c-4aca-9c8b-b8f5d7dc9f04 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.819282] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Skipping network cache update for instance because it is Building. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 528.819439] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Skipping network cache update for instance because it is Building. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 528.819865] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Skipping network cache update for instance because it is Building. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 528.819865] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Skipping network cache update for instance because it is Building. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 528.819865] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Skipping network cache update for instance because it is Building. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 528.819995] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Didn't find any instances for network info cache update. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 528.820704] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 906184f9-27b9-4ce5-aa35-d2a4cdaecae3 could not be found. [ 528.822633] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 528.822633] env[61243]: INFO nova.compute.manager [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 528.822633] env[61243]: DEBUG oslo.service.loopingcall [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 528.822633] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.822633] env[61243]: DEBUG nova.compute.manager [-] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 528.822633] env[61243]: DEBUG nova.network.neutron [-] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 528.823555] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.823784] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.824361] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.824567] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.824758] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.824920] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61243) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 528.825080] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager.update_available_resource {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.886016] env[61243]: DEBUG nova.network.neutron [-] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.915343] env[61243]: DEBUG oslo_concurrency.lockutils [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.916128] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 528.916416] env[61243]: DEBUG oslo_concurrency.lockutils [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.916572] env[61243]: DEBUG oslo_concurrency.lockutils [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.916989] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 528.917622] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d3093e50-fa5b-4e95-9ca8-f114c29dab29 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.938381] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 528.938381] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 528.938381] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b3534e7-d0dc-4970-88c8-5f62e90da520 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.950098] env[61243]: DEBUG nova.compute.manager [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 528.953172] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a65c29e8-bf2e-4781-aa52-772746477cc5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.969118] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 528.969118] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5252a61f-10c6-4307-827c-4230c2111220" [ 528.969118] env[61243]: _type = "Task" [ 528.969118] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.980731] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Preparing fetch location {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 528.981331] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Creating directory with path [datastore2] vmware_temp/cb51ae20-1ca3-45be-8d5a-cece6a315290/6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 528.981642] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-010076d1-9782-41f2-a1ef-9e89d8eb1715 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.016160] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Created directory with path [datastore2] vmware_temp/cb51ae20-1ca3-45be-8d5a-cece6a315290/6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 529.016160] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Fetch image to [datastore2] vmware_temp/cb51ae20-1ca3-45be-8d5a-cece6a315290/6142e969-c114-4502-aa93-c018fb915a86/tmp-sparse.vmdk {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 529.016160] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Downloading image file data 6142e969-c114-4502-aa93-c018fb915a86 to [datastore2] vmware_temp/cb51ae20-1ca3-45be-8d5a-cece6a315290/6142e969-c114-4502-aa93-c018fb915a86/tmp-sparse.vmdk on the data store datastore2 {{(pid=61243) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 529.019369] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-327898f7-f153-4693-a03a-210427304aa8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.033725] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759cbc6b-909d-499f-bc3f-d67a70fa8dcd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.046099] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-216238ae-c6c7-453e-a4be-2330b470837c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.087760] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2620c570-1158-4ea8-b2b2-97eeb04fcd6e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.095276] env[61243]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-9474e5c7-7300-4304-b2e4-ac5bcbcb352e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.119537] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Downloading image file data 6142e969-c114-4502-aa93-c018fb915a86 to the data store datastore2 {{(pid=61243) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 529.164799] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eaf0bc3-d361-48f0-97a7-3e31872d86cc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.178646] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c682b708-726a-4222-9ef1-fbacc3220953 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.210899] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bbd5e45-10f1-46c2-aabf-34aa567aad28 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.218608] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a210142-5be2-4b7b-b98d-0eb311774bca {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.233365] env[61243]: DEBUG nova.compute.provider_tree [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 529.246965] env[61243]: DEBUG oslo_vmware.rw_handles [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/cb51ae20-1ca3-45be-8d5a-cece6a315290/6142e969-c114-4502-aa93-c018fb915a86/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61243) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 529.331520] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.388966] env[61243]: DEBUG nova.network.neutron [-] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.391672] env[61243]: DEBUG nova.network.neutron [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Successfully created port: 1ece1a4e-33e6-4b29-bda3-868bbf198097 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 529.738178] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 529.748761] env[61243]: DEBUG nova.compute.manager [req-c3596ba9-c470-4ea9-99ab-ab951d4151bd req-b8a2a47c-f094-43a8-9ca8-aad19da2b51d service nova] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Received event network-vif-deleted-9171227f-a895-4cad-ab56-8f231377bcad {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.749115] env[61243]: DEBUG nova.compute.manager [req-c3596ba9-c470-4ea9-99ab-ab951d4151bd req-b8a2a47c-f094-43a8-9ca8-aad19da2b51d service nova] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Received event network-changed-6f27b665-bff9-433b-8b38-7954acda35e2 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.749212] env[61243]: DEBUG nova.compute.manager [req-c3596ba9-c470-4ea9-99ab-ab951d4151bd req-b8a2a47c-f094-43a8-9ca8-aad19da2b51d service nova] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Refreshing instance network info cache due to event network-changed-6f27b665-bff9-433b-8b38-7954acda35e2. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 529.749854] env[61243]: DEBUG oslo_concurrency.lockutils [req-c3596ba9-c470-4ea9-99ab-ab951d4151bd req-b8a2a47c-f094-43a8-9ca8-aad19da2b51d service nova] Acquiring lock "refresh_cache-906184f9-27b9-4ce5-aa35-d2a4cdaecae3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.749854] env[61243]: DEBUG oslo_concurrency.lockutils [req-c3596ba9-c470-4ea9-99ab-ab951d4151bd req-b8a2a47c-f094-43a8-9ca8-aad19da2b51d service nova] Acquired lock "refresh_cache-906184f9-27b9-4ce5-aa35-d2a4cdaecae3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.749854] env[61243]: DEBUG nova.network.neutron [req-c3596ba9-c470-4ea9-99ab-ab951d4151bd req-b8a2a47c-f094-43a8-9ca8-aad19da2b51d service nova] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Refreshing network info cache for port 6f27b665-bff9-433b-8b38-7954acda35e2 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 529.898754] env[61243]: INFO nova.compute.manager [-] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Took 1.08 seconds to deallocate network for instance. [ 529.903683] env[61243]: DEBUG nova.compute.claims [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 529.903683] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.962265] env[61243]: DEBUG nova.compute.manager [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 530.002565] env[61243]: DEBUG nova.virt.hardware [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 530.002825] env[61243]: DEBUG nova.virt.hardware [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 530.002978] env[61243]: DEBUG nova.virt.hardware [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 530.003167] env[61243]: DEBUG nova.virt.hardware [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 530.003310] env[61243]: DEBUG nova.virt.hardware [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 530.003464] env[61243]: DEBUG nova.virt.hardware [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 530.003734] env[61243]: DEBUG nova.virt.hardware [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 530.003917] env[61243]: DEBUG nova.virt.hardware [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 530.004098] env[61243]: DEBUG nova.virt.hardware [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 530.004257] env[61243]: DEBUG nova.virt.hardware [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 530.004441] env[61243]: DEBUG nova.virt.hardware [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 530.008884] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef6024cf-c40c-46e4-82bc-617a6ae26092 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.022959] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383336f7-9d78-4977-9823-72461b03aafa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.089960] env[61243]: DEBUG oslo_vmware.rw_handles [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Completed reading data from the image iterator. {{(pid=61243) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 530.089960] env[61243]: DEBUG oslo_vmware.rw_handles [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/cb51ae20-1ca3-45be-8d5a-cece6a315290/6142e969-c114-4502-aa93-c018fb915a86/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 530.152213] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Downloaded image file data 6142e969-c114-4502-aa93-c018fb915a86 to vmware_temp/cb51ae20-1ca3-45be-8d5a-cece6a315290/6142e969-c114-4502-aa93-c018fb915a86/tmp-sparse.vmdk on the data store datastore2 {{(pid=61243) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 530.154541] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Caching image {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 530.154541] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Copying Virtual Disk [datastore2] vmware_temp/cb51ae20-1ca3-45be-8d5a-cece6a315290/6142e969-c114-4502-aa93-c018fb915a86/tmp-sparse.vmdk to [datastore2] vmware_temp/cb51ae20-1ca3-45be-8d5a-cece6a315290/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 530.154958] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a3a1efb3-2d0a-4fe5-b6ba-54336449463c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.164380] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 530.164380] env[61243]: value = "task-1338685" [ 530.164380] env[61243]: _type = "Task" [ 530.164380] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.174341] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338685, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.210506] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Acquiring lock "8c93ef89-5d27-48d3-8c44-79eb23d66ee3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.210506] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Lock "8c93ef89-5d27-48d3-8c44-79eb23d66ee3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.254234] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.319s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.254757] env[61243]: DEBUG nova.compute.manager [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 530.257397] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.852s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.261527] env[61243]: INFO nova.compute.claims [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 530.282263] env[61243]: DEBUG nova.network.neutron [req-c3596ba9-c470-4ea9-99ab-ab951d4151bd req-b8a2a47c-f094-43a8-9ca8-aad19da2b51d service nova] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 530.421646] env[61243]: DEBUG nova.network.neutron [req-c3596ba9-c470-4ea9-99ab-ab951d4151bd req-b8a2a47c-f094-43a8-9ca8-aad19da2b51d service nova] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.678361] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338685, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.718766] env[61243]: DEBUG nova.compute.manager [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 530.769324] env[61243]: DEBUG nova.compute.utils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 530.773950] env[61243]: DEBUG nova.compute.manager [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 530.776092] env[61243]: DEBUG nova.network.neutron [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 530.926037] env[61243]: DEBUG oslo_concurrency.lockutils [req-c3596ba9-c470-4ea9-99ab-ab951d4151bd req-b8a2a47c-f094-43a8-9ca8-aad19da2b51d service nova] Releasing lock "refresh_cache-906184f9-27b9-4ce5-aa35-d2a4cdaecae3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.089147] env[61243]: DEBUG nova.policy [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c98c7ad8df4042a5aae96380c2dbb8ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d1226c068904a19a9a09c96e86c0567', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 531.179465] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338685, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.685543} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.179758] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Copied Virtual Disk [datastore2] vmware_temp/cb51ae20-1ca3-45be-8d5a-cece6a315290/6142e969-c114-4502-aa93-c018fb915a86/tmp-sparse.vmdk to [datastore2] vmware_temp/cb51ae20-1ca3-45be-8d5a-cece6a315290/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 531.179938] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Deleting the datastore file [datastore2] vmware_temp/cb51ae20-1ca3-45be-8d5a-cece6a315290/6142e969-c114-4502-aa93-c018fb915a86/tmp-sparse.vmdk {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 531.180695] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-edd79dd2-25d3-45e0-88e5-e710c3c28f52 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.191116] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 531.191116] env[61243]: value = "task-1338686" [ 531.191116] env[61243]: _type = "Task" [ 531.191116] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.201586] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338686, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.248944] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.274426] env[61243]: DEBUG nova.compute.manager [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 531.479695] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80c6dd9-5e16-4782-8686-95cabf87dbd1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.489935] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc79de7b-4a1d-4c5c-a7fd-0e0534377973 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.535688] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b52caae-3bdf-4671-be40-8ab3fb177f81 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.545635] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c5361f9-c38d-4ef2-bf73-8b9ee58133b7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.562218] env[61243]: DEBUG nova.compute.provider_tree [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 531.704703] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338686, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025569} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.705279] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 531.705807] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Moving file from [datastore2] vmware_temp/cb51ae20-1ca3-45be-8d5a-cece6a315290/6142e969-c114-4502-aa93-c018fb915a86 to [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86. {{(pid=61243) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 531.705807] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-ac3b0399-c901-415d-bbdf-9c85c5af238f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.715554] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 531.715554] env[61243]: value = "task-1338687" [ 531.715554] env[61243]: _type = "Task" [ 531.715554] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.726445] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338687, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.065016] env[61243]: DEBUG nova.scheduler.client.report [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 532.230749] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338687, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.029118} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.230872] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] File moved {{(pid=61243) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 532.231127] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Cleaning up location [datastore2] vmware_temp/cb51ae20-1ca3-45be-8d5a-cece6a315290 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 532.231457] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Deleting the datastore file [datastore2] vmware_temp/cb51ae20-1ca3-45be-8d5a-cece6a315290 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 532.231768] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fdbd82d8-80ed-4aa6-b760-c14a6ad09c23 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.246015] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 532.246015] env[61243]: value = "task-1338688" [ 532.246015] env[61243]: _type = "Task" [ 532.246015] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.257148] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338688, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.285997] env[61243]: DEBUG nova.compute.manager [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 532.322821] env[61243]: DEBUG nova.virt.hardware [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 532.323195] env[61243]: DEBUG nova.virt.hardware [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 532.323316] env[61243]: DEBUG nova.virt.hardware [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 532.323510] env[61243]: DEBUG nova.virt.hardware [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 532.323659] env[61243]: DEBUG nova.virt.hardware [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 532.323803] env[61243]: DEBUG nova.virt.hardware [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 532.324014] env[61243]: DEBUG nova.virt.hardware [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 532.326023] env[61243]: DEBUG nova.virt.hardware [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 532.326023] env[61243]: DEBUG nova.virt.hardware [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 532.326023] env[61243]: DEBUG nova.virt.hardware [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 532.326023] env[61243]: DEBUG nova.virt.hardware [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 532.326656] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-710a49b6-387e-45ae-b97e-1d61b4570ebb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.335121] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ce2ad0-ba20-4908-a9f8-d087f0d4ee10 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.573069] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.314s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.573069] env[61243]: DEBUG nova.compute.manager [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 532.575925] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.039s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.577530] env[61243]: INFO nova.compute.claims [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 532.699196] env[61243]: DEBUG nova.network.neutron [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Successfully created port: 937400db-3ff6-4e49-971b-479c6d6b0fc8 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 532.758278] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338688, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025911} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.758278] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 532.760318] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ac999bb-cd1d-4908-ac17-4403a5f206bf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.770016] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 532.770016] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5225ebfe-5d48-dc11-03a0-4f2e6c4bf64b" [ 532.770016] env[61243]: _type = "Task" [ 532.770016] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.778948] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5225ebfe-5d48-dc11-03a0-4f2e6c4bf64b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.082765] env[61243]: DEBUG nova.compute.utils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 533.087962] env[61243]: DEBUG nova.compute.manager [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 533.088062] env[61243]: DEBUG nova.network.neutron [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 533.279967] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5225ebfe-5d48-dc11-03a0-4f2e6c4bf64b, 'name': SearchDatastore_Task, 'duration_secs': 0.008948} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.280490] env[61243]: DEBUG oslo_concurrency.lockutils [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.280765] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 1b1b2bac-67af-4921-a49a-429017ef7a3b/1b1b2bac-67af-4921-a49a-429017ef7a3b.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 533.281046] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8f2bab6-3366-4ea9-8142-3fe9d5e3601e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.289763] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 533.289763] env[61243]: value = "task-1338689" [ 533.289763] env[61243]: _type = "Task" [ 533.289763] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.298260] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338689, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.398778] env[61243]: DEBUG nova.policy [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4f62a66ae024c0f88b94b434123207c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22e2848953de459a9fcc2b78391d708c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 533.589449] env[61243]: DEBUG nova.compute.manager [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 533.741160] env[61243]: ERROR nova.compute.manager [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1ece1a4e-33e6-4b29-bda3-868bbf198097, please check neutron logs for more information. [ 533.741160] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 533.741160] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.741160] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 533.741160] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.741160] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 533.741160] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.741160] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 533.741160] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.741160] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 533.741160] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.741160] env[61243]: ERROR nova.compute.manager raise self.value [ 533.741160] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.741160] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 533.741160] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.741160] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 533.741630] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.741630] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 533.741630] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1ece1a4e-33e6-4b29-bda3-868bbf198097, please check neutron logs for more information. [ 533.741630] env[61243]: ERROR nova.compute.manager [ 533.742419] env[61243]: Traceback (most recent call last): [ 533.742591] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 533.742591] env[61243]: listener.cb(fileno) [ 533.742676] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.742676] env[61243]: result = function(*args, **kwargs) [ 533.743377] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 533.743377] env[61243]: return func(*args, **kwargs) [ 533.743377] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.743377] env[61243]: raise e [ 533.743377] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.743377] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 533.743377] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.743377] env[61243]: created_port_ids = self._update_ports_for_instance( [ 533.743377] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.743377] env[61243]: with excutils.save_and_reraise_exception(): [ 533.743377] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.743377] env[61243]: self.force_reraise() [ 533.743377] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.743377] env[61243]: raise self.value [ 533.743377] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.743377] env[61243]: updated_port = self._update_port( [ 533.743377] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.743377] env[61243]: _ensure_no_port_binding_failure(port) [ 533.743377] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.743377] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 533.743377] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 1ece1a4e-33e6-4b29-bda3-868bbf198097, please check neutron logs for more information. [ 533.743377] env[61243]: Removing descriptor: 16 [ 533.746773] env[61243]: ERROR nova.compute.manager [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1ece1a4e-33e6-4b29-bda3-868bbf198097, please check neutron logs for more information. [ 533.746773] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Traceback (most recent call last): [ 533.746773] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 533.746773] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] yield resources [ 533.746773] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 533.746773] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] self.driver.spawn(context, instance, image_meta, [ 533.746773] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 533.746773] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 533.746773] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 533.746773] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] vm_ref = self.build_virtual_machine(instance, [ 533.746773] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 533.747101] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] vif_infos = vmwarevif.get_vif_info(self._session, [ 533.747101] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 533.747101] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] for vif in network_info: [ 533.747101] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 533.747101] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] return self._sync_wrapper(fn, *args, **kwargs) [ 533.747101] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 533.747101] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] self.wait() [ 533.747101] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 533.747101] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] self[:] = self._gt.wait() [ 533.747101] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 533.747101] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] return self._exit_event.wait() [ 533.747101] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 533.747101] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] result = hub.switch() [ 533.747451] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 533.747451] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] return self.greenlet.switch() [ 533.747451] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.747451] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] result = function(*args, **kwargs) [ 533.747451] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 533.747451] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] return func(*args, **kwargs) [ 533.747451] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.747451] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] raise e [ 533.747451] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.747451] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] nwinfo = self.network_api.allocate_for_instance( [ 533.747451] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.747451] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] created_port_ids = self._update_ports_for_instance( [ 533.747451] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.747978] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] with excutils.save_and_reraise_exception(): [ 533.747978] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.747978] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] self.force_reraise() [ 533.747978] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.747978] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] raise self.value [ 533.747978] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.747978] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] updated_port = self._update_port( [ 533.747978] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.747978] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] _ensure_no_port_binding_failure(port) [ 533.747978] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.747978] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] raise exception.PortBindingFailed(port_id=port['id']) [ 533.747978] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] nova.exception.PortBindingFailed: Binding failed for port 1ece1a4e-33e6-4b29-bda3-868bbf198097, please check neutron logs for more information. [ 533.747978] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] [ 533.748434] env[61243]: INFO nova.compute.manager [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Terminating instance [ 533.749273] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Acquiring lock "refresh_cache-ffae890f-e876-4382-a21e-35b7e9c05da9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.749589] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Acquired lock "refresh_cache-ffae890f-e876-4382-a21e-35b7e9c05da9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.749767] env[61243]: DEBUG nova.network.neutron [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 533.809169] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338689, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.839057] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f7e199-2ca8-40c7-b646-5705a45537c1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.847969] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1ef143-440a-4455-81bf-66d4d16111ff {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.887903] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ac148f-2c0f-410f-9905-2aa2fcbf6efe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.897446] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfdfbeb8-17e2-49c2-a2e7-86ff44fa8d93 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.915733] env[61243]: DEBUG nova.compute.provider_tree [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 534.302268] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338689, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.766381} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.303052] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 1b1b2bac-67af-4921-a49a-429017ef7a3b/1b1b2bac-67af-4921-a49a-429017ef7a3b.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 534.303052] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 534.303052] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-03727f40-65fc-4264-bfbf-efc8b8380618 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.309895] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 534.309895] env[61243]: value = "task-1338690" [ 534.309895] env[61243]: _type = "Task" [ 534.309895] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.317833] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338690, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.416242] env[61243]: DEBUG nova.network.neutron [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.418675] env[61243]: DEBUG nova.scheduler.client.report [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 534.460129] env[61243]: DEBUG nova.compute.manager [req-4c1ab10b-f6a3-4855-8a10-da62df1b8849 req-a9b5775c-04ab-4e24-a789-32d9a5998de9 service nova] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Received event network-vif-deleted-6f27b665-bff9-433b-8b38-7954acda35e2 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 534.603867] env[61243]: DEBUG nova.compute.manager [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 534.609094] env[61243]: DEBUG nova.network.neutron [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.628702] env[61243]: DEBUG nova.virt.hardware [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 534.628955] env[61243]: DEBUG nova.virt.hardware [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 534.629125] env[61243]: DEBUG nova.virt.hardware [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 534.629336] env[61243]: DEBUG nova.virt.hardware [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 534.629493] env[61243]: DEBUG nova.virt.hardware [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 534.629639] env[61243]: DEBUG nova.virt.hardware [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 534.630607] env[61243]: DEBUG nova.virt.hardware [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 534.631253] env[61243]: DEBUG nova.virt.hardware [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 534.631534] env[61243]: DEBUG nova.virt.hardware [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 534.631710] env[61243]: DEBUG nova.virt.hardware [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 534.631897] env[61243]: DEBUG nova.virt.hardware [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 534.634379] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8d89b2-b605-47a1-b608-07a688056164 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.647753] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a15256-29e7-4949-b67f-33d2f028234a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.824467] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338690, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06341} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.825094] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 534.825801] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb9c0d40-f6d7-4333-a454-7510a6309de4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.856662] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 1b1b2bac-67af-4921-a49a-429017ef7a3b/1b1b2bac-67af-4921-a49a-429017ef7a3b.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 534.857036] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fbd3c142-298e-436a-b7be-a757fba12f80 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.878512] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 534.878512] env[61243]: value = "task-1338691" [ 534.878512] env[61243]: _type = "Task" [ 534.878512] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.883720] env[61243]: DEBUG nova.network.neutron [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Successfully created port: 5847ee11-06ae-4d8d-bba0-419f07abff41 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 534.891401] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338691, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.925213] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.349s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.925658] env[61243]: DEBUG nova.compute.manager [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 534.928843] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 7.920s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.112633] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Releasing lock "refresh_cache-ffae890f-e876-4382-a21e-35b7e9c05da9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.114605] env[61243]: DEBUG nova.compute.manager [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 535.114605] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 535.114605] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ff3c7ea3-fc38-48d1-9c41-61879f5c890c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.126325] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6733b2-2871-421d-9373-0acd467cfcf9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.155809] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ffae890f-e876-4382-a21e-35b7e9c05da9 could not be found. [ 535.156236] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 535.156527] env[61243]: INFO nova.compute.manager [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 535.160158] env[61243]: DEBUG oslo.service.loopingcall [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 535.160415] env[61243]: DEBUG nova.compute.manager [-] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 535.160540] env[61243]: DEBUG nova.network.neutron [-] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 535.219983] env[61243]: DEBUG nova.network.neutron [-] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 535.396427] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338691, 'name': ReconfigVM_Task, 'duration_secs': 0.361952} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.396702] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 1b1b2bac-67af-4921-a49a-429017ef7a3b/1b1b2bac-67af-4921-a49a-429017ef7a3b.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 535.397451] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1f5cf1f7-14f4-424d-91b0-2898a9121738 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.404264] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 535.404264] env[61243]: value = "task-1338692" [ 535.404264] env[61243]: _type = "Task" [ 535.404264] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.418530] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338692, 'name': Rename_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.436198] env[61243]: DEBUG nova.compute.utils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 535.440615] env[61243]: DEBUG nova.compute.manager [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 535.442882] env[61243]: DEBUG nova.network.neutron [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 535.553122] env[61243]: DEBUG nova.policy [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d4b3d0748c14b41898236daf41fc7f3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c05d3c1e23c34a96834cadb2203425ed', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 535.693664] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0a70df-9e7a-4760-b875-1006e26615e9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.705440] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab6080a-e0a5-4b9f-8152-11ed54facb77 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.747154] env[61243]: DEBUG nova.network.neutron [-] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.754603] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-971a3116-e7be-4f21-9c61-b5441c304891 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.769965] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ac2d9a-b24f-450c-969f-4aa08e01d39f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.792733] env[61243]: DEBUG nova.compute.provider_tree [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 535.917923] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338692, 'name': Rename_Task, 'duration_secs': 0.137796} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.917923] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 535.917923] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb4f363e-6e7c-43b3-9b5f-93f58fb2b901 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.930196] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 535.930196] env[61243]: value = "task-1338693" [ 535.930196] env[61243]: _type = "Task" [ 535.930196] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.943787] env[61243]: DEBUG nova.compute.manager [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 535.948125] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338693, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.189324] env[61243]: DEBUG oslo_concurrency.lockutils [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Acquiring lock "84a32e46-ccf7-47ca-986c-8c59d5bf43a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.189440] env[61243]: DEBUG oslo_concurrency.lockutils [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Lock "84a32e46-ccf7-47ca-986c-8c59d5bf43a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.258529] env[61243]: INFO nova.compute.manager [-] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Took 1.10 seconds to deallocate network for instance. [ 536.260841] env[61243]: DEBUG nova.compute.claims [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 536.261032] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.317712] env[61243]: ERROR nova.scheduler.client.report [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [req-cf9f22e4-ddc0-4fdc-b57e-7baf49d0bde7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-cf9f22e4-ddc0-4fdc-b57e-7baf49d0bde7"}]}: nova.exception.PortBindingFailed: Binding failed for port 9171227f-a895-4cad-ab56-8f231377bcad, please check neutron logs for more information. [ 536.341376] env[61243]: DEBUG nova.scheduler.client.report [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 536.359867] env[61243]: DEBUG nova.scheduler.client.report [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 536.360279] env[61243]: DEBUG nova.compute.provider_tree [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 536.371999] env[61243]: DEBUG nova.scheduler.client.report [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 536.398916] env[61243]: DEBUG nova.scheduler.client.report [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 536.440934] env[61243]: DEBUG oslo_vmware.api [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338693, 'name': PowerOnVM_Task, 'duration_secs': 0.449503} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.443099] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 536.443328] env[61243]: INFO nova.compute.manager [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Took 8.73 seconds to spawn the instance on the hypervisor. [ 536.443572] env[61243]: DEBUG nova.compute.manager [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 536.447523] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ff39c1-2893-4388-9189-f9688b69a0d2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.640502] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8dc84fe-fd16-49fd-88c8-d1a9377c0f2d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.648503] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ada5b4-7ff7-4da8-8f96-d7958db2054e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.684747] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da6ab6d-3f8d-41ff-9490-e6baf1ec7666 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.695537] env[61243]: DEBUG nova.compute.manager [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 536.703244] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0913675-7fd3-45e7-ab2d-2b3a6e141dc3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.725228] env[61243]: DEBUG nova.compute.provider_tree [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 536.959065] env[61243]: DEBUG nova.compute.manager [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 536.971274] env[61243]: INFO nova.compute.manager [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Took 16.73 seconds to build instance. [ 536.998825] env[61243]: DEBUG nova.virt.hardware [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 536.998825] env[61243]: DEBUG nova.virt.hardware [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 536.998825] env[61243]: DEBUG nova.virt.hardware [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.999058] env[61243]: DEBUG nova.virt.hardware [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 536.999099] env[61243]: DEBUG nova.virt.hardware [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.999262] env[61243]: DEBUG nova.virt.hardware [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 536.999620] env[61243]: DEBUG nova.virt.hardware [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 536.999620] env[61243]: DEBUG nova.virt.hardware [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 536.999722] env[61243]: DEBUG nova.virt.hardware [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 536.999887] env[61243]: DEBUG nova.virt.hardware [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 537.001017] env[61243]: DEBUG nova.virt.hardware [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 537.001234] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb6c8506-9a68-493a-a21b-fdabccc2e01b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.010490] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908621b7-f124-4d82-940d-6024ddea9032 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.170025] env[61243]: DEBUG nova.network.neutron [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Successfully created port: d044cd3a-6023-4270-aa0d-08b4b677f8f3 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 537.238438] env[61243]: DEBUG oslo_concurrency.lockutils [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.264350] env[61243]: DEBUG nova.scheduler.client.report [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 14 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 537.264598] env[61243]: DEBUG nova.compute.provider_tree [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 14 to 15 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 537.264777] env[61243]: DEBUG nova.compute.provider_tree [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 537.474862] env[61243]: DEBUG oslo_concurrency.lockutils [None req-67c9f5bd-9dc3-4910-a3e6-f49612e46435 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Lock "1b1b2bac-67af-4921-a49a-429017ef7a3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.246s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.615750] env[61243]: ERROR nova.compute.manager [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 937400db-3ff6-4e49-971b-479c6d6b0fc8, please check neutron logs for more information. [ 537.615750] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 537.615750] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.615750] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 537.615750] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.615750] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 537.615750] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.615750] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 537.615750] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.615750] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 537.615750] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.615750] env[61243]: ERROR nova.compute.manager raise self.value [ 537.615750] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.615750] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 537.615750] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.615750] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 537.616364] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.616364] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 537.616364] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 937400db-3ff6-4e49-971b-479c6d6b0fc8, please check neutron logs for more information. [ 537.616364] env[61243]: ERROR nova.compute.manager [ 537.616364] env[61243]: Traceback (most recent call last): [ 537.616364] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 537.616364] env[61243]: listener.cb(fileno) [ 537.616364] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.616364] env[61243]: result = function(*args, **kwargs) [ 537.616364] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 537.616364] env[61243]: return func(*args, **kwargs) [ 537.616364] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.616364] env[61243]: raise e [ 537.616364] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.616364] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 537.616364] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.616364] env[61243]: created_port_ids = self._update_ports_for_instance( [ 537.616364] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.616364] env[61243]: with excutils.save_and_reraise_exception(): [ 537.616364] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.616364] env[61243]: self.force_reraise() [ 537.616364] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.616364] env[61243]: raise self.value [ 537.616364] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.616364] env[61243]: updated_port = self._update_port( [ 537.616364] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.616364] env[61243]: _ensure_no_port_binding_failure(port) [ 537.616364] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.616364] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 537.617527] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 937400db-3ff6-4e49-971b-479c6d6b0fc8, please check neutron logs for more information. [ 537.617527] env[61243]: Removing descriptor: 15 [ 537.617527] env[61243]: ERROR nova.compute.manager [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 937400db-3ff6-4e49-971b-479c6d6b0fc8, please check neutron logs for more information. [ 537.617527] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Traceback (most recent call last): [ 537.617527] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 537.617527] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] yield resources [ 537.617527] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 537.617527] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] self.driver.spawn(context, instance, image_meta, [ 537.617527] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 537.617527] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] self._vmops.spawn(context, instance, image_meta, injected_files, [ 537.617527] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 537.617527] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] vm_ref = self.build_virtual_machine(instance, [ 537.618138] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 537.618138] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] vif_infos = vmwarevif.get_vif_info(self._session, [ 537.618138] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 537.618138] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] for vif in network_info: [ 537.618138] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 537.618138] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] return self._sync_wrapper(fn, *args, **kwargs) [ 537.618138] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 537.618138] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] self.wait() [ 537.618138] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 537.618138] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] self[:] = self._gt.wait() [ 537.618138] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 537.618138] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] return self._exit_event.wait() [ 537.618138] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 537.618547] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] result = hub.switch() [ 537.618547] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 537.618547] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] return self.greenlet.switch() [ 537.618547] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.618547] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] result = function(*args, **kwargs) [ 537.618547] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 537.618547] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] return func(*args, **kwargs) [ 537.618547] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.618547] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] raise e [ 537.618547] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.618547] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] nwinfo = self.network_api.allocate_for_instance( [ 537.618547] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.618547] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] created_port_ids = self._update_ports_for_instance( [ 537.618949] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.618949] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] with excutils.save_and_reraise_exception(): [ 537.618949] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.618949] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] self.force_reraise() [ 537.618949] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.618949] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] raise self.value [ 537.618949] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.618949] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] updated_port = self._update_port( [ 537.618949] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.618949] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] _ensure_no_port_binding_failure(port) [ 537.618949] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.618949] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] raise exception.PortBindingFailed(port_id=port['id']) [ 537.619372] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] nova.exception.PortBindingFailed: Binding failed for port 937400db-3ff6-4e49-971b-479c6d6b0fc8, please check neutron logs for more information. [ 537.619372] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] [ 537.619372] env[61243]: INFO nova.compute.manager [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Terminating instance [ 537.619372] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Acquiring lock "refresh_cache-6c4e205a-6f9d-4439-add7-376fe718d189" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.619372] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Acquired lock "refresh_cache-6c4e205a-6f9d-4439-add7-376fe718d189" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.619508] env[61243]: DEBUG nova.network.neutron [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 537.771855] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.843s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.772966] env[61243]: ERROR nova.compute.manager [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9171227f-a895-4cad-ab56-8f231377bcad, please check neutron logs for more information. [ 537.772966] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Traceback (most recent call last): [ 537.772966] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 537.772966] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] self.driver.spawn(context, instance, image_meta, [ 537.772966] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 537.772966] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 537.772966] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 537.772966] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] vm_ref = self.build_virtual_machine(instance, [ 537.772966] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 537.772966] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] vif_infos = vmwarevif.get_vif_info(self._session, [ 537.772966] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 537.773437] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] for vif in network_info: [ 537.773437] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 537.773437] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] return self._sync_wrapper(fn, *args, **kwargs) [ 537.773437] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 537.773437] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] self.wait() [ 537.773437] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 537.773437] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] self[:] = self._gt.wait() [ 537.773437] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 537.773437] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] return self._exit_event.wait() [ 537.773437] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 537.773437] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] result = hub.switch() [ 537.773437] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 537.773437] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] return self.greenlet.switch() [ 537.773943] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.773943] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] result = function(*args, **kwargs) [ 537.773943] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 537.773943] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] return func(*args, **kwargs) [ 537.773943] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.773943] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] raise e [ 537.773943] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.773943] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] nwinfo = self.network_api.allocate_for_instance( [ 537.773943] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.773943] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] created_port_ids = self._update_ports_for_instance( [ 537.773943] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.773943] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] with excutils.save_and_reraise_exception(): [ 537.773943] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.774380] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] self.force_reraise() [ 537.774380] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.774380] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] raise self.value [ 537.774380] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.774380] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] updated_port = self._update_port( [ 537.774380] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.774380] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] _ensure_no_port_binding_failure(port) [ 537.774380] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.774380] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] raise exception.PortBindingFailed(port_id=port['id']) [ 537.774380] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] nova.exception.PortBindingFailed: Binding failed for port 9171227f-a895-4cad-ab56-8f231377bcad, please check neutron logs for more information. [ 537.774380] env[61243]: ERROR nova.compute.manager [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] [ 537.774881] env[61243]: DEBUG nova.compute.utils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Binding failed for port 9171227f-a895-4cad-ab56-8f231377bcad, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 537.777368] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.214s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.779020] env[61243]: INFO nova.compute.claims [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 537.788260] env[61243]: DEBUG nova.compute.manager [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Build of instance 35aeb5b2-ca5f-457b-9545-05ae609e5ac9 was re-scheduled: Binding failed for port 9171227f-a895-4cad-ab56-8f231377bcad, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 537.788260] env[61243]: DEBUG nova.compute.manager [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 537.788260] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Acquiring lock "refresh_cache-35aeb5b2-ca5f-457b-9545-05ae609e5ac9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.788260] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Acquired lock "refresh_cache-35aeb5b2-ca5f-457b-9545-05ae609e5ac9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.788556] env[61243]: DEBUG nova.network.neutron [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 538.155083] env[61243]: DEBUG nova.network.neutron [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.259531] env[61243]: DEBUG oslo_concurrency.lockutils [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Acquiring lock "e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.259904] env[61243]: DEBUG oslo_concurrency.lockutils [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Lock "e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.276750] env[61243]: DEBUG nova.network.neutron [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.326453] env[61243]: DEBUG nova.network.neutron [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.473059] env[61243]: DEBUG nova.network.neutron [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.762919] env[61243]: DEBUG nova.compute.manager [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 538.780529] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Releasing lock "refresh_cache-6c4e205a-6f9d-4439-add7-376fe718d189" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.781534] env[61243]: DEBUG nova.compute.manager [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 538.781534] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 538.782018] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3de86d1d-7a27-46c7-a635-fb21907348fd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.798476] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fbfdf1a-8c19-4b3b-a016-d32679bfd771 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.837502] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6c4e205a-6f9d-4439-add7-376fe718d189 could not be found. [ 538.837502] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 538.838175] env[61243]: INFO nova.compute.manager [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Took 0.06 seconds to destroy the instance on the hypervisor. [ 538.838482] env[61243]: DEBUG oslo.service.loopingcall [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 538.842461] env[61243]: DEBUG nova.compute.manager [-] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 538.842588] env[61243]: DEBUG nova.network.neutron [-] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 538.895560] env[61243]: DEBUG nova.network.neutron [-] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.977627] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Releasing lock "refresh_cache-35aeb5b2-ca5f-457b-9545-05ae609e5ac9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.977887] env[61243]: DEBUG nova.compute.manager [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 538.978095] env[61243]: DEBUG nova.compute.manager [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 538.978191] env[61243]: DEBUG nova.network.neutron [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 539.008564] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b545d8a5-780c-4691-9bcd-bf1b8f6e839e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.012245] env[61243]: DEBUG nova.network.neutron [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.018852] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756e7f7c-a1a8-41e3-a936-8170537764ed {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.050363] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454b0deb-e7d9-4c13-9b06-f9690f88d2b0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.060112] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d041165-5d36-441e-86a9-a22ead1ac5ce {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.075823] env[61243]: DEBUG nova.compute.provider_tree [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 539.234220] env[61243]: ERROR nova.compute.manager [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5847ee11-06ae-4d8d-bba0-419f07abff41, please check neutron logs for more information. [ 539.234220] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 539.234220] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.234220] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 539.234220] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 539.234220] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 539.234220] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 539.234220] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 539.234220] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.234220] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 539.234220] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.234220] env[61243]: ERROR nova.compute.manager raise self.value [ 539.234220] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 539.234220] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 539.234220] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.234220] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 539.234635] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.234635] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 539.234635] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5847ee11-06ae-4d8d-bba0-419f07abff41, please check neutron logs for more information. [ 539.234635] env[61243]: ERROR nova.compute.manager [ 539.234635] env[61243]: Traceback (most recent call last): [ 539.234635] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 539.234635] env[61243]: listener.cb(fileno) [ 539.234635] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.234635] env[61243]: result = function(*args, **kwargs) [ 539.234635] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 539.234635] env[61243]: return func(*args, **kwargs) [ 539.234635] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.234635] env[61243]: raise e [ 539.234635] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.234635] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 539.234635] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 539.234635] env[61243]: created_port_ids = self._update_ports_for_instance( [ 539.234635] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 539.234635] env[61243]: with excutils.save_and_reraise_exception(): [ 539.234635] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.234635] env[61243]: self.force_reraise() [ 539.234635] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.234635] env[61243]: raise self.value [ 539.234635] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 539.234635] env[61243]: updated_port = self._update_port( [ 539.234635] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.234635] env[61243]: _ensure_no_port_binding_failure(port) [ 539.234635] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.234635] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 539.235330] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 5847ee11-06ae-4d8d-bba0-419f07abff41, please check neutron logs for more information. [ 539.235330] env[61243]: Removing descriptor: 17 [ 539.235330] env[61243]: ERROR nova.compute.manager [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5847ee11-06ae-4d8d-bba0-419f07abff41, please check neutron logs for more information. [ 539.235330] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Traceback (most recent call last): [ 539.235330] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 539.235330] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] yield resources [ 539.235330] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 539.235330] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] self.driver.spawn(context, instance, image_meta, [ 539.235330] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 539.235330] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] self._vmops.spawn(context, instance, image_meta, injected_files, [ 539.235330] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 539.235330] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] vm_ref = self.build_virtual_machine(instance, [ 539.235625] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 539.235625] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] vif_infos = vmwarevif.get_vif_info(self._session, [ 539.235625] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 539.235625] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] for vif in network_info: [ 539.235625] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 539.235625] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] return self._sync_wrapper(fn, *args, **kwargs) [ 539.235625] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 539.235625] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] self.wait() [ 539.235625] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 539.235625] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] self[:] = self._gt.wait() [ 539.235625] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 539.235625] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] return self._exit_event.wait() [ 539.235625] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 539.235937] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] result = hub.switch() [ 539.235937] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 539.235937] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] return self.greenlet.switch() [ 539.235937] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.235937] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] result = function(*args, **kwargs) [ 539.235937] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 539.235937] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] return func(*args, **kwargs) [ 539.235937] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.235937] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] raise e [ 539.235937] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.235937] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] nwinfo = self.network_api.allocate_for_instance( [ 539.235937] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 539.235937] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] created_port_ids = self._update_ports_for_instance( [ 539.240333] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 539.240333] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] with excutils.save_and_reraise_exception(): [ 539.240333] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.240333] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] self.force_reraise() [ 539.240333] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.240333] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] raise self.value [ 539.240333] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 539.240333] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] updated_port = self._update_port( [ 539.240333] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.240333] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] _ensure_no_port_binding_failure(port) [ 539.240333] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.240333] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] raise exception.PortBindingFailed(port_id=port['id']) [ 539.240726] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] nova.exception.PortBindingFailed: Binding failed for port 5847ee11-06ae-4d8d-bba0-419f07abff41, please check neutron logs for more information. [ 539.240726] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] [ 539.240726] env[61243]: INFO nova.compute.manager [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Terminating instance [ 539.240726] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Acquiring lock "refresh_cache-da977758-95d5-4fb6-aaa3-7b832404b527" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.240726] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Acquired lock "refresh_cache-da977758-95d5-4fb6-aaa3-7b832404b527" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.240726] env[61243]: DEBUG nova.network.neutron [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 539.287789] env[61243]: DEBUG oslo_concurrency.lockutils [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.364848] env[61243]: DEBUG nova.compute.manager [req-f6773088-7849-4490-be6f-2aa6deb6a463 req-8a786e19-3af7-4485-bf27-55c374f06f17 service nova] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Received event network-changed-1ece1a4e-33e6-4b29-bda3-868bbf198097 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.366574] env[61243]: DEBUG nova.compute.manager [req-f6773088-7849-4490-be6f-2aa6deb6a463 req-8a786e19-3af7-4485-bf27-55c374f06f17 service nova] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Refreshing instance network info cache due to event network-changed-1ece1a4e-33e6-4b29-bda3-868bbf198097. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 539.366574] env[61243]: DEBUG oslo_concurrency.lockutils [req-f6773088-7849-4490-be6f-2aa6deb6a463 req-8a786e19-3af7-4485-bf27-55c374f06f17 service nova] Acquiring lock "refresh_cache-ffae890f-e876-4382-a21e-35b7e9c05da9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.366574] env[61243]: DEBUG oslo_concurrency.lockutils [req-f6773088-7849-4490-be6f-2aa6deb6a463 req-8a786e19-3af7-4485-bf27-55c374f06f17 service nova] Acquired lock "refresh_cache-ffae890f-e876-4382-a21e-35b7e9c05da9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.366574] env[61243]: DEBUG nova.network.neutron [req-f6773088-7849-4490-be6f-2aa6deb6a463 req-8a786e19-3af7-4485-bf27-55c374f06f17 service nova] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Refreshing network info cache for port 1ece1a4e-33e6-4b29-bda3-868bbf198097 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 539.407791] env[61243]: DEBUG nova.network.neutron [-] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.515800] env[61243]: DEBUG nova.network.neutron [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.579276] env[61243]: DEBUG nova.scheduler.client.report [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 539.708537] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.708537] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.777134] env[61243]: DEBUG nova.network.neutron [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.911085] env[61243]: INFO nova.compute.manager [-] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Took 1.07 seconds to deallocate network for instance. [ 539.914673] env[61243]: DEBUG nova.compute.claims [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 539.914673] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.928840] env[61243]: DEBUG nova.network.neutron [req-f6773088-7849-4490-be6f-2aa6deb6a463 req-8a786e19-3af7-4485-bf27-55c374f06f17 service nova] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.991772] env[61243]: DEBUG nova.network.neutron [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.019945] env[61243]: INFO nova.compute.manager [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] [instance: 35aeb5b2-ca5f-457b-9545-05ae609e5ac9] Took 1.04 seconds to deallocate network for instance. [ 540.087030] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.307s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.087030] env[61243]: DEBUG nova.compute.manager [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 540.091068] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 10.760s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.091444] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.091708] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61243) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 540.092356] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.189s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.095929] env[61243]: DEBUG nova.network.neutron [req-f6773088-7849-4490-be6f-2aa6deb6a463 req-8a786e19-3af7-4485-bf27-55c374f06f17 service nova] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.098665] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c0b2b62-dae6-461c-b552-b891ac5f6555 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.106749] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc95bb6-3a60-4812-bfa9-104d3eecdf43 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.125847] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99db19c7-e565-4439-9354-59e1bb576fee {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.137420] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8279bcd9-b936-4691-a565-885fa1900666 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.173469] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181508MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=61243) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 540.173624] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.200861] env[61243]: ERROR nova.compute.manager [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d044cd3a-6023-4270-aa0d-08b4b677f8f3, please check neutron logs for more information. [ 540.200861] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 540.200861] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.200861] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 540.200861] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.200861] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 540.200861] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.200861] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 540.200861] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.200861] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 540.200861] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.200861] env[61243]: ERROR nova.compute.manager raise self.value [ 540.200861] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.200861] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 540.200861] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.200861] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 540.201322] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.201322] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 540.201322] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d044cd3a-6023-4270-aa0d-08b4b677f8f3, please check neutron logs for more information. [ 540.201322] env[61243]: ERROR nova.compute.manager [ 540.201322] env[61243]: Traceback (most recent call last): [ 540.201322] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 540.201322] env[61243]: listener.cb(fileno) [ 540.201322] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.201322] env[61243]: result = function(*args, **kwargs) [ 540.201322] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 540.201322] env[61243]: return func(*args, **kwargs) [ 540.201322] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.201322] env[61243]: raise e [ 540.201322] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.201322] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 540.201322] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.201322] env[61243]: created_port_ids = self._update_ports_for_instance( [ 540.201322] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.201322] env[61243]: with excutils.save_and_reraise_exception(): [ 540.201322] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.201322] env[61243]: self.force_reraise() [ 540.201322] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.201322] env[61243]: raise self.value [ 540.201322] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.201322] env[61243]: updated_port = self._update_port( [ 540.201322] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.201322] env[61243]: _ensure_no_port_binding_failure(port) [ 540.201322] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.201322] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 540.203068] env[61243]: nova.exception.PortBindingFailed: Binding failed for port d044cd3a-6023-4270-aa0d-08b4b677f8f3, please check neutron logs for more information. [ 540.203068] env[61243]: Removing descriptor: 16 [ 540.203068] env[61243]: ERROR nova.compute.manager [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d044cd3a-6023-4270-aa0d-08b4b677f8f3, please check neutron logs for more information. [ 540.203068] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Traceback (most recent call last): [ 540.203068] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 540.203068] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] yield resources [ 540.203068] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 540.203068] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] self.driver.spawn(context, instance, image_meta, [ 540.203068] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 540.203068] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.203068] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.203068] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] vm_ref = self.build_virtual_machine(instance, [ 540.203483] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.203483] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.203483] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.203483] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] for vif in network_info: [ 540.203483] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.203483] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] return self._sync_wrapper(fn, *args, **kwargs) [ 540.203483] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.203483] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] self.wait() [ 540.203483] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.203483] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] self[:] = self._gt.wait() [ 540.203483] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.203483] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] return self._exit_event.wait() [ 540.203483] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.203810] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] result = hub.switch() [ 540.203810] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.203810] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] return self.greenlet.switch() [ 540.203810] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.203810] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] result = function(*args, **kwargs) [ 540.203810] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 540.203810] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] return func(*args, **kwargs) [ 540.203810] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.203810] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] raise e [ 540.203810] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.203810] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] nwinfo = self.network_api.allocate_for_instance( [ 540.203810] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.203810] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] created_port_ids = self._update_ports_for_instance( [ 540.204166] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.204166] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] with excutils.save_and_reraise_exception(): [ 540.204166] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.204166] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] self.force_reraise() [ 540.204166] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.204166] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] raise self.value [ 540.204166] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.204166] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] updated_port = self._update_port( [ 540.204166] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.204166] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] _ensure_no_port_binding_failure(port) [ 540.204166] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.204166] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] raise exception.PortBindingFailed(port_id=port['id']) [ 540.204443] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] nova.exception.PortBindingFailed: Binding failed for port d044cd3a-6023-4270-aa0d-08b4b677f8f3, please check neutron logs for more information. [ 540.204443] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] [ 540.204443] env[61243]: INFO nova.compute.manager [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Terminating instance [ 540.205769] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Acquiring lock "refresh_cache-f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.207608] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Acquired lock "refresh_cache-f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.207608] env[61243]: DEBUG nova.network.neutron [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 540.496223] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Releasing lock "refresh_cache-da977758-95d5-4fb6-aaa3-7b832404b527" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.496223] env[61243]: DEBUG nova.compute.manager [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 540.496223] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 540.496223] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4349e7b5-7f41-4920-b362-349539678c22 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.506788] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-276ac1c6-1a35-4501-8c73-978fa0cad8ae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.534379] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance da977758-95d5-4fb6-aaa3-7b832404b527 could not be found. [ 540.534621] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 540.534807] env[61243]: INFO nova.compute.manager [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Took 0.04 seconds to destroy the instance on the hypervisor. [ 540.535052] env[61243]: DEBUG oslo.service.loopingcall [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 540.535466] env[61243]: DEBUG nova.compute.manager [-] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 540.535562] env[61243]: DEBUG nova.network.neutron [-] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 540.592430] env[61243]: DEBUG nova.compute.utils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 540.593750] env[61243]: DEBUG nova.compute.manager [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 540.593923] env[61243]: DEBUG nova.network.neutron [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 540.607971] env[61243]: DEBUG oslo_concurrency.lockutils [req-f6773088-7849-4490-be6f-2aa6deb6a463 req-8a786e19-3af7-4485-bf27-55c374f06f17 service nova] Releasing lock "refresh_cache-ffae890f-e876-4382-a21e-35b7e9c05da9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.607971] env[61243]: DEBUG nova.compute.manager [req-f6773088-7849-4490-be6f-2aa6deb6a463 req-8a786e19-3af7-4485-bf27-55c374f06f17 service nova] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Received event network-vif-deleted-1ece1a4e-33e6-4b29-bda3-868bbf198097 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 540.694107] env[61243]: DEBUG nova.network.neutron [-] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.763195] env[61243]: DEBUG nova.network.neutron [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.807060] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-585da058-68d6-419c-b85f-9ce577bc468d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.816339] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db6a56f-2ca1-496c-9180-52f87d7e9849 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.849490] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8cb6952-2230-4ece-aadd-61ecbbcde4be {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.861407] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce92cf3-6eba-4a04-b265-5e9b58c1d78e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.878592] env[61243]: INFO nova.compute.manager [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Rebuilding instance [ 540.882924] env[61243]: DEBUG nova.compute.provider_tree [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 540.886081] env[61243]: DEBUG nova.policy [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3fbcab418aa54e6cb6021a029c3c83a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '01fcdb23e8ed4448ba6a0b3a190fbfff', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 540.947693] env[61243]: DEBUG nova.compute.manager [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 540.948534] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae35f220-91a6-4f11-ab6f-f35f4c07164f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.063942] env[61243]: INFO nova.scheduler.client.report [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Deleted allocations for instance 35aeb5b2-ca5f-457b-9545-05ae609e5ac9 [ 541.094992] env[61243]: DEBUG nova.network.neutron [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.102841] env[61243]: DEBUG nova.compute.manager [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 541.195959] env[61243]: DEBUG nova.network.neutron [-] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.409233] env[61243]: ERROR nova.scheduler.client.report [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [req-67f34030-9542-4e9e-8be6-2c4ebad4c16c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-67f34030-9542-4e9e-8be6-2c4ebad4c16c"}]}: nova.exception.PortBindingFailed: Binding failed for port 6f27b665-bff9-433b-8b38-7954acda35e2, please check neutron logs for more information. [ 541.429704] env[61243]: DEBUG nova.scheduler.client.report [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 541.448343] env[61243]: DEBUG nova.scheduler.client.report [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 541.448343] env[61243]: DEBUG nova.compute.provider_tree [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 541.460846] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 541.462113] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-488ab57b-a586-4e59-81f0-ed8f4578d468 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.467299] env[61243]: DEBUG nova.scheduler.client.report [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 541.473872] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 541.473872] env[61243]: value = "task-1338699" [ 541.473872] env[61243]: _type = "Task" [ 541.473872] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.483346] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338699, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.501445] env[61243]: DEBUG nova.scheduler.client.report [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 541.574297] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e6109b7-81e1-4c66-bcee-b7fc3e7e841f tempest-ServerDiagnosticsTest-2018084012 tempest-ServerDiagnosticsTest-2018084012-project-member] Lock "35aeb5b2-ca5f-457b-9545-05ae609e5ac9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.684s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.607837] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Releasing lock "refresh_cache-f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.608304] env[61243]: DEBUG nova.compute.manager [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 541.608758] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 541.616730] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-00654fb6-32f2-4812-96c8-748c945cc9ba {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.632305] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3454da6d-54ea-49eb-a5a5-d521d5b7ce22 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.663795] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d could not be found. [ 541.664080] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 541.664293] env[61243]: INFO nova.compute.manager [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Took 0.06 seconds to destroy the instance on the hypervisor. [ 541.664553] env[61243]: DEBUG oslo.service.loopingcall [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 541.671085] env[61243]: DEBUG nova.compute.manager [-] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 541.671085] env[61243]: DEBUG nova.network.neutron [-] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 541.700048] env[61243]: INFO nova.compute.manager [-] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Took 1.16 seconds to deallocate network for instance. [ 541.704333] env[61243]: DEBUG nova.compute.claims [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 541.704881] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.743901] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900ccfb6-85f0-48ec-b506-533aed50cf3b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.749369] env[61243]: DEBUG nova.network.neutron [-] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 541.762724] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328f7585-81d1-4ab5-bb38-48e4a2d5a34b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.805012] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a4f995b-41bc-46d7-bd1d-80aac2f85246 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.816432] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e45eac2-e84a-4e86-a04f-f0c6243ab5c0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.837358] env[61243]: DEBUG nova.compute.provider_tree [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 541.990636] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338699, 'name': PowerOffVM_Task, 'duration_secs': 0.137605} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.991082] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 541.993716] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 541.993716] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac14ad0-b7f9-414f-a27b-12c36dd40711 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.003853] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 542.005163] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6f38c8c3-884e-4a69-b90a-1ef53189150f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.030998] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 542.032683] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 542.033122] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Deleting the datastore file [datastore2] 1b1b2bac-67af-4921-a49a-429017ef7a3b {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 542.033468] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df740c69-a111-48c0-9d03-ae88b626572e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.041795] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 542.041795] env[61243]: value = "task-1338702" [ 542.041795] env[61243]: _type = "Task" [ 542.041795] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.051379] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Acquiring lock "762cbca7-4729-458a-b2af-2978723a25dc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.051842] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Lock "762cbca7-4729-458a-b2af-2978723a25dc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.056456] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338702, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.080530] env[61243]: DEBUG nova.compute.manager [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 542.124519] env[61243]: DEBUG nova.compute.manager [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 542.161757] env[61243]: DEBUG nova.virt.hardware [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 542.162439] env[61243]: DEBUG nova.virt.hardware [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 542.164334] env[61243]: DEBUG nova.virt.hardware [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 542.164605] env[61243]: DEBUG nova.virt.hardware [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 542.164803] env[61243]: DEBUG nova.virt.hardware [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 542.165010] env[61243]: DEBUG nova.virt.hardware [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 542.165301] env[61243]: DEBUG nova.virt.hardware [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 542.165504] env[61243]: DEBUG nova.virt.hardware [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 542.166091] env[61243]: DEBUG nova.virt.hardware [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 542.166091] env[61243]: DEBUG nova.virt.hardware [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 542.166232] env[61243]: DEBUG nova.virt.hardware [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 542.169512] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-170a1fad-e7a0-45c4-8ff2-d85941986f1e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.179301] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3fd3588-67bd-4210-9a7b-78af9fbb4812 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.255706] env[61243]: DEBUG nova.network.neutron [-] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.366993] env[61243]: ERROR nova.scheduler.client.report [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [req-c99f085e-f1df-4b0d-a100-7ea4809a5d3f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c99f085e-f1df-4b0d-a100-7ea4809a5d3f"}]}: nova.exception.PortBindingFailed: Binding failed for port 6f27b665-bff9-433b-8b38-7954acda35e2, please check neutron logs for more information. [ 542.395532] env[61243]: DEBUG nova.scheduler.client.report [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 542.412808] env[61243]: DEBUG nova.scheduler.client.report [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 542.414149] env[61243]: DEBUG nova.compute.provider_tree [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 542.431358] env[61243]: DEBUG nova.scheduler.client.report [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 542.459176] env[61243]: DEBUG nova.scheduler.client.report [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 542.556457] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338702, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149343} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.556457] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 542.556457] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 542.556457] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 542.612650] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.633124] env[61243]: DEBUG nova.network.neutron [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Successfully created port: b2bb5756-8638-4396-acc4-0b333a9578c9 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 542.669078] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a1eeeb9-3f6b-4194-bd16-1357e09fd3f1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.679020] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06059364-b929-4fbb-9db2-cd01ab920878 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.714911] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8576c6-9953-47b7-8b99-b877bc286541 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.723663] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-562a42c7-ad44-44d1-878c-4ec0642806b9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.739176] env[61243]: DEBUG nova.compute.provider_tree [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 542.762332] env[61243]: INFO nova.compute.manager [-] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Took 1.09 seconds to deallocate network for instance. [ 542.765095] env[61243]: DEBUG nova.compute.claims [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 542.765310] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.279430] env[61243]: DEBUG nova.compute.manager [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Received event network-changed-937400db-3ff6-4e49-971b-479c6d6b0fc8 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 543.283826] env[61243]: DEBUG nova.compute.manager [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Refreshing instance network info cache due to event network-changed-937400db-3ff6-4e49-971b-479c6d6b0fc8. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 543.283826] env[61243]: DEBUG oslo_concurrency.lockutils [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] Acquiring lock "refresh_cache-6c4e205a-6f9d-4439-add7-376fe718d189" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.283826] env[61243]: DEBUG oslo_concurrency.lockutils [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] Acquired lock "refresh_cache-6c4e205a-6f9d-4439-add7-376fe718d189" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.283826] env[61243]: DEBUG nova.network.neutron [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Refreshing network info cache for port 937400db-3ff6-4e49-971b-479c6d6b0fc8 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 543.293138] env[61243]: DEBUG nova.scheduler.client.report [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 18 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 543.293138] env[61243]: DEBUG nova.compute.provider_tree [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 18 to 19 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 543.293138] env[61243]: DEBUG nova.compute.provider_tree [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 543.604147] env[61243]: DEBUG nova.virt.hardware [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 543.605016] env[61243]: DEBUG nova.virt.hardware [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 543.605016] env[61243]: DEBUG nova.virt.hardware [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 543.605016] env[61243]: DEBUG nova.virt.hardware [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 543.607023] env[61243]: DEBUG nova.virt.hardware [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 543.607023] env[61243]: DEBUG nova.virt.hardware [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 543.607023] env[61243]: DEBUG nova.virt.hardware [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 543.607023] env[61243]: DEBUG nova.virt.hardware [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 543.607023] env[61243]: DEBUG nova.virt.hardware [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 543.607223] env[61243]: DEBUG nova.virt.hardware [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 543.607223] env[61243]: DEBUG nova.virt.hardware [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 543.607845] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dba8407-1bb3-493c-9795-8793300ec7d8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.618868] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f56601cc-fa4e-4bf6-a9ea-4c07fbccf768 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.636882] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Instance VIF info [] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 543.644630] env[61243]: DEBUG oslo.service.loopingcall [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 543.648018] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 543.648018] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3a262767-0636-4297-a5d7-cb5961fc69b6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.668615] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 543.668615] env[61243]: value = "task-1338703" [ 543.668615] env[61243]: _type = "Task" [ 543.668615] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.682275] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338703, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.797080] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 3.705s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.798272] env[61243]: ERROR nova.compute.manager [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6f27b665-bff9-433b-8b38-7954acda35e2, please check neutron logs for more information. [ 543.798272] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Traceback (most recent call last): [ 543.798272] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 543.798272] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] self.driver.spawn(context, instance, image_meta, [ 543.798272] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 543.798272] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 543.798272] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 543.798272] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] vm_ref = self.build_virtual_machine(instance, [ 543.798272] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 543.798272] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] vif_infos = vmwarevif.get_vif_info(self._session, [ 543.798272] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 543.801931] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] for vif in network_info: [ 543.801931] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 543.801931] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] return self._sync_wrapper(fn, *args, **kwargs) [ 543.801931] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 543.801931] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] self.wait() [ 543.801931] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 543.801931] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] self[:] = self._gt.wait() [ 543.801931] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 543.801931] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] return self._exit_event.wait() [ 543.801931] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 543.801931] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] result = hub.switch() [ 543.801931] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 543.801931] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] return self.greenlet.switch() [ 543.802284] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.802284] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] result = function(*args, **kwargs) [ 543.802284] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 543.802284] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] return func(*args, **kwargs) [ 543.802284] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.802284] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] raise e [ 543.802284] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.802284] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] nwinfo = self.network_api.allocate_for_instance( [ 543.802284] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.802284] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] created_port_ids = self._update_ports_for_instance( [ 543.802284] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.802284] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] with excutils.save_and_reraise_exception(): [ 543.802284] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.803815] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] self.force_reraise() [ 543.803815] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.803815] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] raise self.value [ 543.803815] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.803815] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] updated_port = self._update_port( [ 543.803815] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.803815] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] _ensure_no_port_binding_failure(port) [ 543.803815] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.803815] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] raise exception.PortBindingFailed(port_id=port['id']) [ 543.803815] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] nova.exception.PortBindingFailed: Binding failed for port 6f27b665-bff9-433b-8b38-7954acda35e2, please check neutron logs for more information. [ 543.803815] env[61243]: ERROR nova.compute.manager [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] [ 543.804328] env[61243]: DEBUG nova.compute.utils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Binding failed for port 6f27b665-bff9-433b-8b38-7954acda35e2, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 543.804328] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.556s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.805912] env[61243]: INFO nova.compute.claims [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 543.810043] env[61243]: DEBUG nova.compute.manager [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Build of instance 906184f9-27b9-4ce5-aa35-d2a4cdaecae3 was re-scheduled: Binding failed for port 6f27b665-bff9-433b-8b38-7954acda35e2, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 543.811429] env[61243]: DEBUG nova.compute.manager [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 543.811775] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Acquiring lock "refresh_cache-906184f9-27b9-4ce5-aa35-d2a4cdaecae3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.812554] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Acquired lock "refresh_cache-906184f9-27b9-4ce5-aa35-d2a4cdaecae3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.812919] env[61243]: DEBUG nova.network.neutron [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 543.860122] env[61243]: DEBUG nova.network.neutron [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.068595] env[61243]: DEBUG nova.network.neutron [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.185102] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338703, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.367710] env[61243]: DEBUG nova.network.neutron [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.518380] env[61243]: DEBUG nova.network.neutron [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.576800] env[61243]: DEBUG oslo_concurrency.lockutils [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] Releasing lock "refresh_cache-6c4e205a-6f9d-4439-add7-376fe718d189" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.576939] env[61243]: DEBUG nova.compute.manager [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Received event network-vif-deleted-937400db-3ff6-4e49-971b-479c6d6b0fc8 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 544.577594] env[61243]: DEBUG nova.compute.manager [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Received event network-changed-5847ee11-06ae-4d8d-bba0-419f07abff41 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 544.577594] env[61243]: DEBUG nova.compute.manager [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Refreshing instance network info cache due to event network-changed-5847ee11-06ae-4d8d-bba0-419f07abff41. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 544.577700] env[61243]: DEBUG oslo_concurrency.lockutils [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] Acquiring lock "refresh_cache-da977758-95d5-4fb6-aaa3-7b832404b527" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.577774] env[61243]: DEBUG oslo_concurrency.lockutils [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] Acquired lock "refresh_cache-da977758-95d5-4fb6-aaa3-7b832404b527" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.577925] env[61243]: DEBUG nova.network.neutron [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Refreshing network info cache for port 5847ee11-06ae-4d8d-bba0-419f07abff41 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 544.681265] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338703, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.021492] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Releasing lock "refresh_cache-906184f9-27b9-4ce5-aa35-d2a4cdaecae3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.021724] env[61243]: DEBUG nova.compute.manager [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 545.021893] env[61243]: DEBUG nova.compute.manager [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 545.022105] env[61243]: DEBUG nova.network.neutron [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 545.075585] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57514cec-0333-4ed2-ba4b-24d9bdb1fb3a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.082898] env[61243]: DEBUG nova.network.neutron [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.092356] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d2d816-568f-4085-9b27-256362cda9af {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.135422] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d8bc80-856d-41d6-a02f-971343bc5f23 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.146110] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf6eec4-a80b-469d-9b6a-019f37ef1ed5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.151420] env[61243]: DEBUG nova.network.neutron [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.166198] env[61243]: DEBUG nova.compute.provider_tree [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.184613] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338703, 'name': CreateVM_Task, 'duration_secs': 1.305714} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.185211] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 545.185211] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.185363] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.188411] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 545.188411] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a79153e-bcb9-4298-a32e-3a3e49abe922 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.197349] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 545.197349] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521ea19c-1ad1-106f-babd-f3a77bb83d17" [ 545.197349] env[61243]: _type = "Task" [ 545.197349] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.206710] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521ea19c-1ad1-106f-babd-f3a77bb83d17, 'name': SearchDatastore_Task, 'duration_secs': 0.008571} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.206710] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.206710] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 545.207058] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.207058] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.207152] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 545.207734] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7485485b-2a0c-4b51-a681-6512a9915ef9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.215382] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 545.215614] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 545.216340] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7b3dbcd-a8c3-427a-96cb-44873a5e8c25 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.225381] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 545.225381] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52858981-234d-43ca-4eb4-2d12736140b2" [ 545.225381] env[61243]: _type = "Task" [ 545.225381] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.233891] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52858981-234d-43ca-4eb4-2d12736140b2, 'name': SearchDatastore_Task, 'duration_secs': 0.00804} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.234668] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd36656f-e3cd-4204-be99-550b78442b52 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.244605] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 545.244605] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f8cb18-377b-11df-0ddd-71214e2dfb8e" [ 545.244605] env[61243]: _type = "Task" [ 545.244605] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.255033] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f8cb18-377b-11df-0ddd-71214e2dfb8e, 'name': SearchDatastore_Task, 'duration_secs': 0.008449} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.255033] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.255033] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 1b1b2bac-67af-4921-a49a-429017ef7a3b/1b1b2bac-67af-4921-a49a-429017ef7a3b.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 545.256373] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4d0d9915-698b-40ab-8e7b-a2d9debb9ce8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.264224] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 545.264224] env[61243]: value = "task-1338705" [ 545.264224] env[61243]: _type = "Task" [ 545.264224] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.274555] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338705, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.369758] env[61243]: DEBUG nova.network.neutron [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.589169] env[61243]: DEBUG nova.network.neutron [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.670324] env[61243]: DEBUG nova.scheduler.client.report [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 545.781305] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338705, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.511627} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.781730] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 1b1b2bac-67af-4921-a49a-429017ef7a3b/1b1b2bac-67af-4921-a49a-429017ef7a3b.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 545.782059] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 545.782697] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-99ca2518-81ff-486a-9652-b1f0a710648b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.789591] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 545.789591] env[61243]: value = "task-1338706" [ 545.789591] env[61243]: _type = "Task" [ 545.789591] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.800081] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338706, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.872543] env[61243]: DEBUG oslo_concurrency.lockutils [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] Releasing lock "refresh_cache-da977758-95d5-4fb6-aaa3-7b832404b527" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.872982] env[61243]: DEBUG nova.compute.manager [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Received event network-vif-deleted-5847ee11-06ae-4d8d-bba0-419f07abff41 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 545.873673] env[61243]: DEBUG nova.compute.manager [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Received event network-changed-d044cd3a-6023-4270-aa0d-08b4b677f8f3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 545.873987] env[61243]: DEBUG nova.compute.manager [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Refreshing instance network info cache due to event network-changed-d044cd3a-6023-4270-aa0d-08b4b677f8f3. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 545.874314] env[61243]: DEBUG oslo_concurrency.lockutils [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] Acquiring lock "refresh_cache-f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.875592] env[61243]: DEBUG oslo_concurrency.lockutils [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] Acquired lock "refresh_cache-f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.875592] env[61243]: DEBUG nova.network.neutron [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Refreshing network info cache for port d044cd3a-6023-4270-aa0d-08b4b677f8f3 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 546.094581] env[61243]: INFO nova.compute.manager [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] [instance: 906184f9-27b9-4ce5-aa35-d2a4cdaecae3] Took 1.07 seconds to deallocate network for instance. [ 546.176524] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.176777] env[61243]: DEBUG nova.compute.manager [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 546.185084] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.920s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.302474] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338706, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06856} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.302934] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 546.306017] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c980c24-58d5-4229-93db-7265985dedce {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.327274] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 1b1b2bac-67af-4921-a49a-429017ef7a3b/1b1b2bac-67af-4921-a49a-429017ef7a3b.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 546.327746] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-043ea206-da1f-461c-874b-beed7cb42eae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.350061] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 546.350061] env[61243]: value = "task-1338708" [ 546.350061] env[61243]: _type = "Task" [ 546.350061] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.358465] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338708, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.413956] env[61243]: DEBUG nova.network.neutron [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.637810] env[61243]: DEBUG nova.network.neutron [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.691342] env[61243]: DEBUG nova.compute.utils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 546.698528] env[61243]: DEBUG nova.compute.manager [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 546.700856] env[61243]: DEBUG nova.network.neutron [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 546.858940] env[61243]: DEBUG nova.policy [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7cff8e952ce1400681b834157eaf9dd6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8fbcee03fda4bff9730bf5c2a97570c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 546.867187] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338708, 'name': ReconfigVM_Task, 'duration_secs': 0.315609} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.870646] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 1b1b2bac-67af-4921-a49a-429017ef7a3b/1b1b2bac-67af-4921-a49a-429017ef7a3b.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 546.872316] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9097baae-4db6-4b5e-b4cd-ac323fe4f4e3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.881332] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 546.881332] env[61243]: value = "task-1338709" [ 546.881332] env[61243]: _type = "Task" [ 546.881332] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.894790] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338709, 'name': Rename_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.966033] env[61243]: ERROR nova.compute.manager [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b2bb5756-8638-4396-acc4-0b333a9578c9, please check neutron logs for more information. [ 546.966033] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 546.966033] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 546.966033] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 546.966033] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.966033] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 546.966033] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.966033] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 546.966033] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.966033] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 546.966033] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.966033] env[61243]: ERROR nova.compute.manager raise self.value [ 546.966033] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.966033] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 546.966033] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.966033] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 546.966528] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.966528] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 546.966528] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b2bb5756-8638-4396-acc4-0b333a9578c9, please check neutron logs for more information. [ 546.966528] env[61243]: ERROR nova.compute.manager [ 546.966528] env[61243]: Traceback (most recent call last): [ 546.966528] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 546.966528] env[61243]: listener.cb(fileno) [ 546.966528] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.966528] env[61243]: result = function(*args, **kwargs) [ 546.966528] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 546.966528] env[61243]: return func(*args, **kwargs) [ 546.966528] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 546.966528] env[61243]: raise e [ 546.966528] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 546.966528] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 546.966528] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.966528] env[61243]: created_port_ids = self._update_ports_for_instance( [ 546.966528] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.966528] env[61243]: with excutils.save_and_reraise_exception(): [ 546.966528] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.966528] env[61243]: self.force_reraise() [ 546.966528] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.966528] env[61243]: raise self.value [ 546.966528] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.966528] env[61243]: updated_port = self._update_port( [ 546.966528] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.966528] env[61243]: _ensure_no_port_binding_failure(port) [ 546.966528] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.966528] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 546.967322] env[61243]: nova.exception.PortBindingFailed: Binding failed for port b2bb5756-8638-4396-acc4-0b333a9578c9, please check neutron logs for more information. [ 546.967322] env[61243]: Removing descriptor: 15 [ 546.967322] env[61243]: ERROR nova.compute.manager [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b2bb5756-8638-4396-acc4-0b333a9578c9, please check neutron logs for more information. [ 546.967322] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Traceback (most recent call last): [ 546.967322] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 546.967322] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] yield resources [ 546.967322] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 546.967322] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] self.driver.spawn(context, instance, image_meta, [ 546.967322] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 546.967322] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] self._vmops.spawn(context, instance, image_meta, injected_files, [ 546.967322] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 546.967322] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] vm_ref = self.build_virtual_machine(instance, [ 546.967663] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 546.967663] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] vif_infos = vmwarevif.get_vif_info(self._session, [ 546.967663] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 546.967663] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] for vif in network_info: [ 546.967663] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 546.967663] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] return self._sync_wrapper(fn, *args, **kwargs) [ 546.967663] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 546.967663] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] self.wait() [ 546.967663] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 546.967663] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] self[:] = self._gt.wait() [ 546.967663] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 546.967663] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] return self._exit_event.wait() [ 546.967663] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 546.967995] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] result = hub.switch() [ 546.967995] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 546.967995] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] return self.greenlet.switch() [ 546.967995] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.967995] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] result = function(*args, **kwargs) [ 546.967995] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 546.967995] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] return func(*args, **kwargs) [ 546.967995] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 546.967995] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] raise e [ 546.967995] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 546.967995] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] nwinfo = self.network_api.allocate_for_instance( [ 546.967995] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.967995] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] created_port_ids = self._update_ports_for_instance( [ 546.968343] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.968343] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] with excutils.save_and_reraise_exception(): [ 546.968343] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.968343] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] self.force_reraise() [ 546.968343] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.968343] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] raise self.value [ 546.968343] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.968343] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] updated_port = self._update_port( [ 546.968343] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.968343] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] _ensure_no_port_binding_failure(port) [ 546.968343] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.968343] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] raise exception.PortBindingFailed(port_id=port['id']) [ 546.968847] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] nova.exception.PortBindingFailed: Binding failed for port b2bb5756-8638-4396-acc4-0b333a9578c9, please check neutron logs for more information. [ 546.968847] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] [ 546.968847] env[61243]: INFO nova.compute.manager [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Terminating instance [ 546.971025] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Acquiring lock "refresh_cache-25a896cb-fd91-4f16-a8cd-6de58fcfcc27" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.971264] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Acquired lock "refresh_cache-25a896cb-fd91-4f16-a8cd-6de58fcfcc27" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.971383] env[61243]: DEBUG nova.network.neutron [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 546.991140] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc98f314-b000-4a17-9b6f-f81dd0ce3557 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.002831] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcc14004-78c6-43f1-b853-f84f531407cf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.039616] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-562673c0-45af-44f9-80fe-cf736b03346f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.047608] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a063f5a2-80fa-4c92-ab2a-875998b45b54 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.063748] env[61243]: DEBUG nova.compute.provider_tree [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 547.145038] env[61243]: INFO nova.scheduler.client.report [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Deleted allocations for instance 906184f9-27b9-4ce5-aa35-d2a4cdaecae3 [ 547.151929] env[61243]: DEBUG oslo_concurrency.lockutils [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] Releasing lock "refresh_cache-f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.151929] env[61243]: DEBUG nova.compute.manager [req-3906d721-07ed-4241-87f6-917ed5fe4811 req-3b8da95e-b132-4307-808e-5d33ffd2e880 service nova] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Received event network-vif-deleted-d044cd3a-6023-4270-aa0d-08b4b677f8f3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 547.195855] env[61243]: DEBUG nova.compute.manager [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 547.392535] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338709, 'name': Rename_Task, 'duration_secs': 0.157351} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.392535] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 547.392736] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f7aebd98-b33d-4a80-800a-3d3207f1a908 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.400251] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 547.400251] env[61243]: value = "task-1338710" [ 547.400251] env[61243]: _type = "Task" [ 547.400251] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.414308] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338710, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.537631] env[61243]: DEBUG nova.network.neutron [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.605953] env[61243]: ERROR nova.scheduler.client.report [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [req-f8df4d92-d53f-42a6-ac3f-5ee0f9e18845] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f8df4d92-d53f-42a6-ac3f-5ee0f9e18845"}]}: nova.exception.PortBindingFailed: Binding failed for port 1ece1a4e-33e6-4b29-bda3-868bbf198097, please check neutron logs for more information. [ 547.622665] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Acquiring lock "3700deef-a438-4920-b35f-8fba370a5984" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.623755] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Lock "3700deef-a438-4920-b35f-8fba370a5984" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.630114] env[61243]: DEBUG nova.scheduler.client.report [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 547.646025] env[61243]: DEBUG nova.scheduler.client.report [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 547.646229] env[61243]: DEBUG nova.compute.provider_tree [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 547.652758] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5573dda3-6252-4522-b0e8-9724da97923f tempest-TenantUsagesTestJSON-1904698900 tempest-TenantUsagesTestJSON-1904698900-project-member] Lock "906184f9-27b9-4ce5-aa35-d2a4cdaecae3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.782s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.665955] env[61243]: DEBUG nova.scheduler.client.report [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 547.693592] env[61243]: DEBUG nova.scheduler.client.report [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 547.738470] env[61243]: DEBUG nova.network.neutron [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.917826] env[61243]: DEBUG oslo_vmware.api [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338710, 'name': PowerOnVM_Task, 'duration_secs': 0.476255} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.918167] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 547.918409] env[61243]: DEBUG nova.compute.manager [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 547.919229] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5cc97b2-13c7-4103-bd69-7f3bf2d1a8b8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.927022] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68aee564-d381-4211-a190-44baa6239428 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.942331] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f352d38-ca49-4145-9b93-37eae2215a94 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.976030] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5322fda5-e3d5-468c-a426-09429c179c8c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.984017] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624ad29b-19df-47c3-b590-5f4c53dad742 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.997769] env[61243]: DEBUG nova.compute.provider_tree [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 548.048635] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Acquiring lock "2451843a-fd32-450e-933e-e635c789834b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.048998] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Lock "2451843a-fd32-450e-933e-e635c789834b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.129932] env[61243]: DEBUG nova.network.neutron [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Successfully created port: 4a04ed28-25ad-4709-a762-c930986c585e {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 548.155311] env[61243]: DEBUG nova.compute.manager [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.208884] env[61243]: DEBUG nova.compute.manager [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 548.243079] env[61243]: DEBUG nova.virt.hardware [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.244610] env[61243]: DEBUG nova.virt.hardware [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.244610] env[61243]: DEBUG nova.virt.hardware [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.244610] env[61243]: DEBUG nova.virt.hardware [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.244610] env[61243]: DEBUG nova.virt.hardware [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.244610] env[61243]: DEBUG nova.virt.hardware [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.244981] env[61243]: DEBUG nova.virt.hardware [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.244981] env[61243]: DEBUG nova.virt.hardware [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.244981] env[61243]: DEBUG nova.virt.hardware [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.245158] env[61243]: DEBUG nova.virt.hardware [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.245515] env[61243]: DEBUG nova.virt.hardware [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 548.246039] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Releasing lock "refresh_cache-25a896cb-fd91-4f16-a8cd-6de58fcfcc27" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.246925] env[61243]: DEBUG nova.compute.manager [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 548.247180] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 548.249109] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3fcdbd7-4c96-4e1d-8145-4fe58f24321d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.254115] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fad9496b-952e-4967-b48f-b50aba09187f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.263882] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427ff1aa-5f58-48f5-91e8-f257d446b625 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.271147] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5668f0ad-90bd-4a04-a24e-c5b1cd75fb21 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.300416] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 25a896cb-fd91-4f16-a8cd-6de58fcfcc27 could not be found. [ 548.300727] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 548.300939] env[61243]: INFO nova.compute.manager [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Took 0.05 seconds to destroy the instance on the hypervisor. [ 548.301644] env[61243]: DEBUG oslo.service.loopingcall [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 548.301644] env[61243]: DEBUG nova.compute.manager [-] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 548.301644] env[61243]: DEBUG nova.network.neutron [-] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 548.340673] env[61243]: DEBUG nova.network.neutron [-] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.446663] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.521258] env[61243]: ERROR nova.scheduler.client.report [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [req-c645e91d-2057-4c8b-a8f7-13bb3d0ea0ce] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c645e91d-2057-4c8b-a8f7-13bb3d0ea0ce"}]}: nova.exception.PortBindingFailed: Binding failed for port 1ece1a4e-33e6-4b29-bda3-868bbf198097, please check neutron logs for more information. [ 548.539506] env[61243]: DEBUG nova.scheduler.client.report [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 548.552659] env[61243]: DEBUG nova.scheduler.client.report [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 548.552882] env[61243]: DEBUG nova.compute.provider_tree [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 22 to 23 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 548.553063] env[61243]: DEBUG nova.compute.provider_tree [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 548.563838] env[61243]: DEBUG nova.scheduler.client.report [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 548.581606] env[61243]: DEBUG nova.scheduler.client.report [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 548.677199] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.798099] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48f2d540-74ae-48ee-b7a2-d0605c1a6b0d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.805549] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bce66994-a8aa-4b77-91c2-1c6ebf4d8e49 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.845327] env[61243]: DEBUG nova.network.neutron [-] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.849635] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a5b207-cc63-4bd0-8845-99a6a27afd07 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.851782] env[61243]: DEBUG nova.compute.manager [req-a4cd1610-c047-4d30-9dee-5fbdab5a9c57 req-1d20d653-01ab-4a13-9c69-cb8255c66fab service nova] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Received event network-changed-b2bb5756-8638-4396-acc4-0b333a9578c9 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 548.851996] env[61243]: DEBUG nova.compute.manager [req-a4cd1610-c047-4d30-9dee-5fbdab5a9c57 req-1d20d653-01ab-4a13-9c69-cb8255c66fab service nova] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Refreshing instance network info cache due to event network-changed-b2bb5756-8638-4396-acc4-0b333a9578c9. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 548.855053] env[61243]: DEBUG oslo_concurrency.lockutils [req-a4cd1610-c047-4d30-9dee-5fbdab5a9c57 req-1d20d653-01ab-4a13-9c69-cb8255c66fab service nova] Acquiring lock "refresh_cache-25a896cb-fd91-4f16-a8cd-6de58fcfcc27" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.855053] env[61243]: DEBUG oslo_concurrency.lockutils [req-a4cd1610-c047-4d30-9dee-5fbdab5a9c57 req-1d20d653-01ab-4a13-9c69-cb8255c66fab service nova] Acquired lock "refresh_cache-25a896cb-fd91-4f16-a8cd-6de58fcfcc27" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.855053] env[61243]: DEBUG nova.network.neutron [req-a4cd1610-c047-4d30-9dee-5fbdab5a9c57 req-1d20d653-01ab-4a13-9c69-cb8255c66fab service nova] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Refreshing network info cache for port b2bb5756-8638-4396-acc4-0b333a9578c9 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 548.861695] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c59017d-2492-433e-8e16-2d0c8fc36856 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.882222] env[61243]: DEBUG nova.compute.provider_tree [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 548.996708] env[61243]: DEBUG oslo_concurrency.lockutils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquiring lock "f21b7977-747e-437f-b03d-8442ed58756e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.996902] env[61243]: DEBUG oslo_concurrency.lockutils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "f21b7977-747e-437f-b03d-8442ed58756e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.062299] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Acquiring lock "e2c498c5-3df1-4b99-a322-a598afd506f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.062736] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Lock "e2c498c5-3df1-4b99-a322-a598afd506f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.354761] env[61243]: INFO nova.compute.manager [-] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Took 1.05 seconds to deallocate network for instance. [ 549.358708] env[61243]: DEBUG nova.compute.claims [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 549.358978] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.398117] env[61243]: DEBUG nova.network.neutron [req-a4cd1610-c047-4d30-9dee-5fbdab5a9c57 req-1d20d653-01ab-4a13-9c69-cb8255c66fab service nova] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.430401] env[61243]: DEBUG nova.scheduler.client.report [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 23 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 549.430669] env[61243]: DEBUG nova.compute.provider_tree [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 23 to 24 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 549.430896] env[61243]: DEBUG nova.compute.provider_tree [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 549.541479] env[61243]: DEBUG nova.network.neutron [req-a4cd1610-c047-4d30-9dee-5fbdab5a9c57 req-1d20d653-01ab-4a13-9c69-cb8255c66fab service nova] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.938906] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 3.757s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.942083] env[61243]: ERROR nova.compute.manager [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1ece1a4e-33e6-4b29-bda3-868bbf198097, please check neutron logs for more information. [ 549.942083] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Traceback (most recent call last): [ 549.942083] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 549.942083] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] self.driver.spawn(context, instance, image_meta, [ 549.942083] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 549.942083] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 549.942083] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 549.942083] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] vm_ref = self.build_virtual_machine(instance, [ 549.942083] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 549.942083] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] vif_infos = vmwarevif.get_vif_info(self._session, [ 549.942083] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 549.942539] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] for vif in network_info: [ 549.942539] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 549.942539] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] return self._sync_wrapper(fn, *args, **kwargs) [ 549.942539] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 549.942539] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] self.wait() [ 549.942539] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 549.942539] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] self[:] = self._gt.wait() [ 549.942539] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 549.942539] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] return self._exit_event.wait() [ 549.942539] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 549.942539] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] result = hub.switch() [ 549.942539] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 549.942539] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] return self.greenlet.switch() [ 549.943137] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.943137] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] result = function(*args, **kwargs) [ 549.943137] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 549.943137] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] return func(*args, **kwargs) [ 549.943137] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 549.943137] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] raise e [ 549.943137] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.943137] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] nwinfo = self.network_api.allocate_for_instance( [ 549.943137] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.943137] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] created_port_ids = self._update_ports_for_instance( [ 549.943137] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.943137] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] with excutils.save_and_reraise_exception(): [ 549.943137] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.944168] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] self.force_reraise() [ 549.944168] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.944168] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] raise self.value [ 549.944168] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.944168] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] updated_port = self._update_port( [ 549.944168] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.944168] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] _ensure_no_port_binding_failure(port) [ 549.944168] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.944168] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] raise exception.PortBindingFailed(port_id=port['id']) [ 549.944168] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] nova.exception.PortBindingFailed: Binding failed for port 1ece1a4e-33e6-4b29-bda3-868bbf198097, please check neutron logs for more information. [ 549.944168] env[61243]: ERROR nova.compute.manager [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] [ 549.944461] env[61243]: DEBUG nova.compute.utils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Binding failed for port 1ece1a4e-33e6-4b29-bda3-868bbf198097, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 549.944461] env[61243]: DEBUG nova.compute.manager [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Build of instance ffae890f-e876-4382-a21e-35b7e9c05da9 was re-scheduled: Binding failed for port 1ece1a4e-33e6-4b29-bda3-868bbf198097, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 549.944461] env[61243]: DEBUG nova.compute.manager [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 549.944571] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Acquiring lock "refresh_cache-ffae890f-e876-4382-a21e-35b7e9c05da9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.946385] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Acquired lock "refresh_cache-ffae890f-e876-4382-a21e-35b7e9c05da9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.946385] env[61243]: DEBUG nova.network.neutron [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 549.946385] env[61243]: DEBUG oslo_concurrency.lockutils [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.708s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.951566] env[61243]: INFO nova.compute.claims [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 549.989185] env[61243]: DEBUG nova.network.neutron [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.045493] env[61243]: DEBUG oslo_concurrency.lockutils [req-a4cd1610-c047-4d30-9dee-5fbdab5a9c57 req-1d20d653-01ab-4a13-9c69-cb8255c66fab service nova] Releasing lock "refresh_cache-25a896cb-fd91-4f16-a8cd-6de58fcfcc27" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.130742] env[61243]: DEBUG nova.network.neutron [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.636458] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Releasing lock "refresh_cache-ffae890f-e876-4382-a21e-35b7e9c05da9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.636458] env[61243]: DEBUG nova.compute.manager [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 550.636458] env[61243]: DEBUG nova.compute.manager [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 550.636458] env[61243]: DEBUG nova.network.neutron [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 550.670481] env[61243]: DEBUG nova.network.neutron [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.065054] env[61243]: DEBUG oslo_concurrency.lockutils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Acquiring lock "5508a717-fb58-49b3-bfd1-b80d76a3c4e1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.065354] env[61243]: DEBUG oslo_concurrency.lockutils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Lock "5508a717-fb58-49b3-bfd1-b80d76a3c4e1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.172518] env[61243]: DEBUG nova.network.neutron [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.268518] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b1683a-7a96-4fd9-9469-81c8ca31d82e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.277600] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3abc28a-57c4-402b-90fe-a8abd29e1a75 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.309501] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08874a2b-f459-4e01-9d68-2dd74fd05f24 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.318385] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71f0f3c-85d4-452c-8e1c-16877100bcf3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.333911] env[61243]: DEBUG nova.compute.provider_tree [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 551.514940] env[61243]: ERROR nova.compute.manager [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4a04ed28-25ad-4709-a762-c930986c585e, please check neutron logs for more information. [ 551.514940] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 551.514940] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.514940] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 551.514940] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.514940] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 551.514940] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.514940] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 551.514940] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.514940] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 551.514940] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.514940] env[61243]: ERROR nova.compute.manager raise self.value [ 551.514940] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.514940] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 551.514940] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.514940] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 551.515372] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.515372] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 551.515372] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4a04ed28-25ad-4709-a762-c930986c585e, please check neutron logs for more information. [ 551.515372] env[61243]: ERROR nova.compute.manager [ 551.515372] env[61243]: Traceback (most recent call last): [ 551.515372] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 551.515372] env[61243]: listener.cb(fileno) [ 551.515372] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.515372] env[61243]: result = function(*args, **kwargs) [ 551.515372] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 551.515372] env[61243]: return func(*args, **kwargs) [ 551.515372] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.515372] env[61243]: raise e [ 551.515372] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.515372] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 551.515372] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.515372] env[61243]: created_port_ids = self._update_ports_for_instance( [ 551.515372] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.515372] env[61243]: with excutils.save_and_reraise_exception(): [ 551.515372] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.515372] env[61243]: self.force_reraise() [ 551.515372] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.515372] env[61243]: raise self.value [ 551.515372] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.515372] env[61243]: updated_port = self._update_port( [ 551.515372] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.515372] env[61243]: _ensure_no_port_binding_failure(port) [ 551.515372] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.515372] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 551.516145] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 4a04ed28-25ad-4709-a762-c930986c585e, please check neutron logs for more information. [ 551.516145] env[61243]: Removing descriptor: 16 [ 551.516145] env[61243]: ERROR nova.compute.manager [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4a04ed28-25ad-4709-a762-c930986c585e, please check neutron logs for more information. [ 551.516145] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Traceback (most recent call last): [ 551.516145] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 551.516145] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] yield resources [ 551.516145] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 551.516145] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] self.driver.spawn(context, instance, image_meta, [ 551.516145] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 551.516145] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.516145] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.516145] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] vm_ref = self.build_virtual_machine(instance, [ 551.516517] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.516517] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.516517] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.516517] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] for vif in network_info: [ 551.516517] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.516517] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] return self._sync_wrapper(fn, *args, **kwargs) [ 551.516517] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.516517] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] self.wait() [ 551.516517] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.516517] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] self[:] = self._gt.wait() [ 551.516517] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.516517] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] return self._exit_event.wait() [ 551.516517] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.516828] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] result = hub.switch() [ 551.516828] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.516828] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] return self.greenlet.switch() [ 551.516828] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.516828] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] result = function(*args, **kwargs) [ 551.516828] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 551.516828] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] return func(*args, **kwargs) [ 551.516828] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.516828] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] raise e [ 551.516828] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.516828] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] nwinfo = self.network_api.allocate_for_instance( [ 551.516828] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.516828] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] created_port_ids = self._update_ports_for_instance( [ 551.517191] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.517191] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] with excutils.save_and_reraise_exception(): [ 551.517191] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.517191] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] self.force_reraise() [ 551.517191] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.517191] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] raise self.value [ 551.517191] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.517191] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] updated_port = self._update_port( [ 551.517191] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.517191] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] _ensure_no_port_binding_failure(port) [ 551.517191] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.517191] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] raise exception.PortBindingFailed(port_id=port['id']) [ 551.517488] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] nova.exception.PortBindingFailed: Binding failed for port 4a04ed28-25ad-4709-a762-c930986c585e, please check neutron logs for more information. [ 551.517488] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] [ 551.517488] env[61243]: INFO nova.compute.manager [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Terminating instance [ 551.520968] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Acquiring lock "refresh_cache-8c93ef89-5d27-48d3-8c44-79eb23d66ee3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.521062] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Acquired lock "refresh_cache-8c93ef89-5d27-48d3-8c44-79eb23d66ee3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.521207] env[61243]: DEBUG nova.network.neutron [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 551.681460] env[61243]: INFO nova.compute.manager [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] [instance: ffae890f-e876-4382-a21e-35b7e9c05da9] Took 1.04 seconds to deallocate network for instance. [ 551.743946] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquiring lock "476632bb-e07c-4660-9ea4-7290884e4761" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.743946] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "476632bb-e07c-4660-9ea4-7290884e4761" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.862824] env[61243]: ERROR nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [req-7d93d315-18b0-4a11-88d7-f569d1f44d7d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7d93d315-18b0-4a11-88d7-f569d1f44d7d"}]} [ 551.882879] env[61243]: DEBUG nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 551.903533] env[61243]: DEBUG nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 551.903835] env[61243]: DEBUG nova.compute.provider_tree [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 551.920298] env[61243]: DEBUG nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 551.944660] env[61243]: DEBUG nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 552.084036] env[61243]: DEBUG nova.network.neutron [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.124363] env[61243]: DEBUG nova.compute.manager [req-db6c7893-2d5c-4462-bab1-fde45fd64cf0 req-dcd4f40d-8b4d-413a-a6cf-0996b70cde46 service nova] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Received event network-vif-deleted-b2bb5756-8638-4396-acc4-0b333a9578c9 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 552.125526] env[61243]: DEBUG nova.compute.manager [req-db6c7893-2d5c-4462-bab1-fde45fd64cf0 req-dcd4f40d-8b4d-413a-a6cf-0996b70cde46 service nova] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Received event network-changed-4a04ed28-25ad-4709-a762-c930986c585e {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 552.125721] env[61243]: DEBUG nova.compute.manager [req-db6c7893-2d5c-4462-bab1-fde45fd64cf0 req-dcd4f40d-8b4d-413a-a6cf-0996b70cde46 service nova] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Refreshing instance network info cache due to event network-changed-4a04ed28-25ad-4709-a762-c930986c585e. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 552.125947] env[61243]: DEBUG oslo_concurrency.lockutils [req-db6c7893-2d5c-4462-bab1-fde45fd64cf0 req-dcd4f40d-8b4d-413a-a6cf-0996b70cde46 service nova] Acquiring lock "refresh_cache-8c93ef89-5d27-48d3-8c44-79eb23d66ee3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.158696] env[61243]: INFO nova.compute.manager [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Rebuilding instance [ 552.233637] env[61243]: DEBUG nova.compute.manager [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 552.234487] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2b9792-e674-45eb-a2b4-e8970a6c2dcf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.285040] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36ee01f8-dafb-42af-96c5-a67368c4f535 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.296011] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-644aff0e-f697-4afc-9334-602ed414660a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.330125] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86cdc41-281c-4a04-85d3-cb48e53c704a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.340773] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9acdc454-cfd0-4197-ad05-45f00eeb1599 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.359847] env[61243]: DEBUG nova.compute.provider_tree [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 552.430187] env[61243]: DEBUG nova.network.neutron [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.720182] env[61243]: INFO nova.scheduler.client.report [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Deleted allocations for instance ffae890f-e876-4382-a21e-35b7e9c05da9 [ 552.747328] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 552.747903] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-45a34d10-9c67-46e6-ba39-7a848a645a84 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.761685] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Waiting for the task: (returnval){ [ 552.761685] env[61243]: value = "task-1338714" [ 552.761685] env[61243]: _type = "Task" [ 552.761685] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.773627] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': task-1338714, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.889998] env[61243]: ERROR nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [req-1a4f2fa4-82d1-4cec-ba98-cd537cd78301] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1a4f2fa4-82d1-4cec-ba98-cd537cd78301"}]} [ 552.911299] env[61243]: DEBUG nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 552.932858] env[61243]: DEBUG nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 552.933371] env[61243]: DEBUG nova.compute.provider_tree [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 552.935595] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Releasing lock "refresh_cache-8c93ef89-5d27-48d3-8c44-79eb23d66ee3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.937175] env[61243]: DEBUG nova.compute.manager [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 552.937175] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 552.937250] env[61243]: DEBUG oslo_concurrency.lockutils [req-db6c7893-2d5c-4462-bab1-fde45fd64cf0 req-dcd4f40d-8b4d-413a-a6cf-0996b70cde46 service nova] Acquired lock "refresh_cache-8c93ef89-5d27-48d3-8c44-79eb23d66ee3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.937430] env[61243]: DEBUG nova.network.neutron [req-db6c7893-2d5c-4462-bab1-fde45fd64cf0 req-dcd4f40d-8b4d-413a-a6cf-0996b70cde46 service nova] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Refreshing network info cache for port 4a04ed28-25ad-4709-a762-c930986c585e {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 552.943018] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6fe969d3-0cc7-44e6-abe4-0d9cf2638d48 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.953215] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8b12e8-9559-4ae2-a8a6-2aa0cafa79e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.970237] env[61243]: DEBUG nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 552.993183] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8c93ef89-5d27-48d3-8c44-79eb23d66ee3 could not be found. [ 552.993183] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 552.993757] env[61243]: INFO nova.compute.manager [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Took 0.06 seconds to destroy the instance on the hypervisor. [ 552.993757] env[61243]: DEBUG oslo.service.loopingcall [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 552.995364] env[61243]: DEBUG nova.compute.manager [-] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 552.995364] env[61243]: DEBUG nova.network.neutron [-] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 553.002342] env[61243]: DEBUG nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 553.022401] env[61243]: DEBUG nova.network.neutron [-] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.159490] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Acquiring lock "b5163d89-ceb8-4c61-ae02-0ae3311b58a9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.159562] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Lock "b5163d89-ceb8-4c61-ae02-0ae3311b58a9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.235722] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e43586-b24c-4276-968b-b37ca7b4a8b5 tempest-ServerDiagnosticsNegativeTest-1956963387 tempest-ServerDiagnosticsNegativeTest-1956963387-project-member] Lock "ffae890f-e876-4382-a21e-35b7e9c05da9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.557s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.291463] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': task-1338714, 'name': PowerOffVM_Task, 'duration_secs': 0.137954} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.292328] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 553.292328] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 553.293106] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05173a2d-4785-4934-8457-a8ebd0f50c48 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.303738] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 553.307486] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97b594f8-c9a1-4d20-8f9e-4eb916525168 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.343598] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 553.344028] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 553.344028] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Deleting the datastore file [datastore2] 1b1b2bac-67af-4921-a49a-429017ef7a3b {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 553.344449] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ccc21228-a216-41b2-859d-e7756747cd8a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.355219] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Waiting for the task: (returnval){ [ 553.355219] env[61243]: value = "task-1338716" [ 553.355219] env[61243]: _type = "Task" [ 553.355219] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.370872] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': task-1338716, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.430438] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a00f668b-3de9-4ade-b95b-67af6df61329 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.444673] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a0bda1-5ae1-4323-8fa0-09c57cd1248a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.494390] env[61243]: DEBUG nova.network.neutron [req-db6c7893-2d5c-4462-bab1-fde45fd64cf0 req-dcd4f40d-8b4d-413a-a6cf-0996b70cde46 service nova] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.498572] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d33a42a-c5d0-4976-bdff-7e1f9095704a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.507238] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7fd9d0-50d4-4f94-9697-d58b8852cc70 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.523318] env[61243]: DEBUG nova.compute.provider_tree [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 553.524753] env[61243]: DEBUG nova.network.neutron [-] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.706230] env[61243]: DEBUG nova.network.neutron [req-db6c7893-2d5c-4462-bab1-fde45fd64cf0 req-dcd4f40d-8b4d-413a-a6cf-0996b70cde46 service nova] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.738503] env[61243]: DEBUG nova.compute.manager [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 553.755141] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquiring lock "461b2ef2-d665-47ab-affa-e525bffe561a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.755779] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "461b2ef2-d665-47ab-affa-e525bffe561a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.876058] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': task-1338716, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101679} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.876058] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 553.876058] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 553.876337] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 554.032898] env[61243]: INFO nova.compute.manager [-] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Took 1.04 seconds to deallocate network for instance. [ 554.036615] env[61243]: DEBUG nova.compute.claims [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 554.036927] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.063853] env[61243]: ERROR nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [req-56fd739e-0a86-441a-9765-aec9f5630049] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-56fd739e-0a86-441a-9765-aec9f5630049"}]} [ 554.090021] env[61243]: DEBUG nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 554.115169] env[61243]: DEBUG nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 554.115169] env[61243]: DEBUG nova.compute.provider_tree [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 554.134963] env[61243]: DEBUG nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 554.161399] env[61243]: DEBUG nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 554.210617] env[61243]: DEBUG oslo_concurrency.lockutils [req-db6c7893-2d5c-4462-bab1-fde45fd64cf0 req-dcd4f40d-8b4d-413a-a6cf-0996b70cde46 service nova] Releasing lock "refresh_cache-8c93ef89-5d27-48d3-8c44-79eb23d66ee3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.263103] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.528792] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-859255c9-4738-42d1-9b47-fdf2b83091b0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.542133] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b809b7-7ae1-40b0-85db-a71bf5acc8b1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.587097] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ddda51f-fb02-4c5c-a896-500a28b011cb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.595965] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1f1bb6-5e32-4e28-8b83-dc56b00810a8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.613512] env[61243]: DEBUG nova.compute.provider_tree [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 554.937667] env[61243]: DEBUG nova.virt.hardware [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 554.937978] env[61243]: DEBUG nova.virt.hardware [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 554.938099] env[61243]: DEBUG nova.virt.hardware [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 554.940624] env[61243]: DEBUG nova.virt.hardware [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 554.940624] env[61243]: DEBUG nova.virt.hardware [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 554.940624] env[61243]: DEBUG nova.virt.hardware [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 554.940624] env[61243]: DEBUG nova.virt.hardware [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 554.940624] env[61243]: DEBUG nova.virt.hardware [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 554.941294] env[61243]: DEBUG nova.virt.hardware [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 554.941294] env[61243]: DEBUG nova.virt.hardware [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 554.941416] env[61243]: DEBUG nova.virt.hardware [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 554.942557] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1fac31-4c6e-4f84-b77c-9246457c32cd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.959106] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19df0864-b2d0-45cd-9083-576831db8be4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.980860] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Instance VIF info [] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 554.989789] env[61243]: DEBUG oslo.service.loopingcall [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 554.989789] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 554.989789] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d8fc64a4-ff79-4cef-839f-0186f8c9710a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.012480] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 555.012480] env[61243]: value = "task-1338717" [ 555.012480] env[61243]: _type = "Task" [ 555.012480] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.023781] env[61243]: DEBUG nova.compute.manager [req-c9195fd5-2bd8-4bcd-aedb-d5bd53c39f07 req-7bdfff88-ff44-4bcf-9876-85378048cb65 service nova] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Received event network-vif-deleted-4a04ed28-25ad-4709-a762-c930986c585e {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 555.032448] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338717, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.146734] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Acquiring lock "b3d98e6f-4b45-4eaf-a79c-b683b32fe053" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.148029] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Lock "b3d98e6f-4b45-4eaf-a79c-b683b32fe053" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.148539] env[61243]: ERROR nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [req-c4e78274-cd79-4d3d-98f1-f41e7834a1d5] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c4e78274-cd79-4d3d-98f1-f41e7834a1d5"}]} [ 555.148982] env[61243]: DEBUG oslo_concurrency.lockutils [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 5.203s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.151152] env[61243]: ERROR nova.compute.manager [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] Failed to build and run instance: nova.exception.ResourceProviderUpdateConflict: A conflict was encountered attempting to update resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 (generation 28): {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c4e78274-cd79-4d3d-98f1-f41e7834a1d5"}]} [ 555.151152] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] Traceback (most recent call last): [ 555.151152] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] File "/opt/stack/nova/nova/compute/manager.py", line 2615, in _build_and_run_instance [ 555.151152] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] with self.rt.instance_claim(context, instance, node, allocs, [ 555.151152] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 555.151152] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] return f(*args, **kwargs) [ 555.151152] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 555.151152] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] self._update(elevated, cn) [ 555.151152] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 555.151471] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] self._update_to_placement(context, compute_node, startup) [ 555.151471] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 555.151471] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 555.151471] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 266, in call [ 555.151471] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] raise attempt.get() [ 555.151471] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 555.151471] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] six.reraise(self.value[0], self.value[1], self.value[2]) [ 555.151471] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 555.151471] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] raise value [ 555.151471] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 555.151471] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 555.151471] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 555.151471] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] self.reportclient.update_from_provider_tree( [ 555.151864] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 555.151864] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] self.set_inventory_for_provider( [ 555.151864] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1003, in set_inventory_for_provider [ 555.151864] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] raise exception.ResourceProviderUpdateConflict( [ 555.151864] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] nova.exception.ResourceProviderUpdateConflict: A conflict was encountered attempting to update resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 (generation 28): {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c4e78274-cd79-4d3d-98f1-f41e7834a1d5"}]} [ 555.151864] env[61243]: ERROR nova.compute.manager [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] [ 555.157947] env[61243]: DEBUG nova.compute.utils [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] A conflict was encountered attempting to update resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 (generation 28): {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource pro {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 555.157947] env[61243]: DEBUG oslo_concurrency.lockutils [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.870s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.158881] env[61243]: INFO nova.compute.claims [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 555.162784] env[61243]: DEBUG nova.compute.manager [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] Build of instance 84a32e46-ccf7-47ca-986c-8c59d5bf43a5 was re-scheduled: A conflict was encountered attempting to update resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 (generation 28): {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c4e78274-cd79-4d3d-98f1-f41e7834a1d5"}]} {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 555.163128] env[61243]: DEBUG nova.compute.manager [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 555.167371] env[61243]: DEBUG oslo_concurrency.lockutils [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Acquiring lock "refresh_cache-84a32e46-ccf7-47ca-986c-8c59d5bf43a5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.167547] env[61243]: DEBUG oslo_concurrency.lockutils [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Acquired lock "refresh_cache-84a32e46-ccf7-47ca-986c-8c59d5bf43a5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.167714] env[61243]: DEBUG nova.network.neutron [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 555.527342] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338717, 'name': CreateVM_Task, 'duration_secs': 0.364091} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.527674] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 555.528268] env[61243]: DEBUG oslo_concurrency.lockutils [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.528424] env[61243]: DEBUG oslo_concurrency.lockutils [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.529459] env[61243]: DEBUG oslo_concurrency.lockutils [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 555.529459] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15af1d0b-a685-40a5-9dd4-0d769e6599e4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.541259] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Waiting for the task: (returnval){ [ 555.541259] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522936f9-eede-fdac-4bd6-07667524687e" [ 555.541259] env[61243]: _type = "Task" [ 555.541259] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.550868] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522936f9-eede-fdac-4bd6-07667524687e, 'name': SearchDatastore_Task, 'duration_secs': 0.009462} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.550868] env[61243]: DEBUG oslo_concurrency.lockutils [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.551058] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 555.551277] env[61243]: DEBUG oslo_concurrency.lockutils [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.551495] env[61243]: DEBUG oslo_concurrency.lockutils [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.552911] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 555.552911] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bfd94f24-e68f-4283-9939-271ca26ac81b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.562786] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 555.563206] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 555.563750] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c015bdd3-93a0-499d-b3f0-17db113ffa65 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.571398] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Waiting for the task: (returnval){ [ 555.571398] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5223b5b8-3250-6fb1-357f-5a6e1d0306bf" [ 555.571398] env[61243]: _type = "Task" [ 555.571398] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.580541] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5223b5b8-3250-6fb1-357f-5a6e1d0306bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.710404] env[61243]: DEBUG nova.network.neutron [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.773610] env[61243]: DEBUG nova.network.neutron [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.085478] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5223b5b8-3250-6fb1-357f-5a6e1d0306bf, 'name': SearchDatastore_Task, 'duration_secs': 0.008995} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.086472] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e71fd20e-ebae-457b-b6b3-8fe4eb9bc76b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.092770] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Waiting for the task: (returnval){ [ 556.092770] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]520ff4d1-8e46-0fd9-c8b2-ae6602bc4f83" [ 556.092770] env[61243]: _type = "Task" [ 556.092770] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.106050] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]520ff4d1-8e46-0fd9-c8b2-ae6602bc4f83, 'name': SearchDatastore_Task, 'duration_secs': 0.009447} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.106494] env[61243]: DEBUG oslo_concurrency.lockutils [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.106635] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 1b1b2bac-67af-4921-a49a-429017ef7a3b/1b1b2bac-67af-4921-a49a-429017ef7a3b.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 556.106897] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ec45ca3f-4c86-4815-b0ea-583dbbc8e970 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.115450] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Waiting for the task: (returnval){ [ 556.115450] env[61243]: value = "task-1338718" [ 556.115450] env[61243]: _type = "Task" [ 556.115450] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.125735] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': task-1338718, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.200300] env[61243]: DEBUG nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 556.214247] env[61243]: DEBUG nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 556.214565] env[61243]: DEBUG nova.compute.provider_tree [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 556.233625] env[61243]: DEBUG nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 556.261165] env[61243]: DEBUG nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 556.282025] env[61243]: DEBUG oslo_concurrency.lockutils [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Releasing lock "refresh_cache-84a32e46-ccf7-47ca-986c-8c59d5bf43a5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.282025] env[61243]: DEBUG nova.compute.manager [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 556.282025] env[61243]: DEBUG nova.compute.manager [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 556.282025] env[61243]: DEBUG nova.network.neutron [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 556.308669] env[61243]: DEBUG nova.network.neutron [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.618591] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e9ea800-af20-4282-8d2a-fe3f06356c13 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.640899] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': task-1338718, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.649611] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26872ccb-3f7f-4832-892b-20227f43c396 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.653425] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Acquiring lock "0b223791-7e95-43fe-bef0-bac6222884ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.654038] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Lock "0b223791-7e95-43fe-bef0-bac6222884ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.686673] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900f291f-cee6-4ef8-8daf-174a3bca5bfd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.694764] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa9976b5-6ca0-42bc-81e5-4c00de4ad234 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.711724] env[61243]: DEBUG nova.compute.provider_tree [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 556.811389] env[61243]: DEBUG nova.network.neutron [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.113197] env[61243]: DEBUG oslo_concurrency.lockutils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Acquiring lock "64efd676-9166-4ffa-a437-6fa6c466ace0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.113472] env[61243]: DEBUG oslo_concurrency.lockutils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Lock "64efd676-9166-4ffa-a437-6fa6c466ace0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.134480] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': task-1338718, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.545657} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.134480] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 1b1b2bac-67af-4921-a49a-429017ef7a3b/1b1b2bac-67af-4921-a49a-429017ef7a3b.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 557.134480] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 557.134480] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7bf6bd22-d492-49aa-8411-356f24a4b189 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.145959] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Waiting for the task: (returnval){ [ 557.145959] env[61243]: value = "task-1338719" [ 557.145959] env[61243]: _type = "Task" [ 557.145959] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.160716] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': task-1338719, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.236906] env[61243]: ERROR nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [req-1f74a17a-dddd-4aad-82ae-0d0f06238652] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1f74a17a-dddd-4aad-82ae-0d0f06238652"}]} [ 557.256393] env[61243]: DEBUG nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 557.289508] env[61243]: DEBUG nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 557.289508] env[61243]: DEBUG nova.compute.provider_tree [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 557.315988] env[61243]: INFO nova.compute.manager [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] [instance: 84a32e46-ccf7-47ca-986c-8c59d5bf43a5] Took 1.04 seconds to deallocate network for instance. [ 557.343779] env[61243]: DEBUG nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 557.344305] env[61243]: DEBUG nova.compute.provider_tree [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 31 to 32 during operation: update_aggregates {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 557.377808] env[61243]: DEBUG nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 557.377808] env[61243]: DEBUG nova.compute.provider_tree [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 32 to 33 during operation: update_traits {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 557.666745] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': task-1338719, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083009} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.667084] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 557.667920] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f26bd3aa-60a0-4659-a0e9-2245e9011297 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.702168] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 1b1b2bac-67af-4921-a49a-429017ef7a3b/1b1b2bac-67af-4921-a49a-429017ef7a3b.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 557.706155] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3af83a5d-ad19-4d57-a51b-3469d31d66d8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.727027] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Waiting for the task: (returnval){ [ 557.727027] env[61243]: value = "task-1338720" [ 557.727027] env[61243]: _type = "Task" [ 557.727027] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.744167] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': task-1338720, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.830831] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Acquiring lock "0625ac10-55b4-488a-aba5-bdaa394045f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.831391] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Lock "0625ac10-55b4-488a-aba5-bdaa394045f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.843877] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Acquiring lock "8efd391b-7b76-4167-bd18-48566271ccc8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.843877] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Lock "8efd391b-7b76-4167-bd18-48566271ccc8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.906242] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2126dcfc-2b24-447f-bbc5-e5c9e0c612a0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.915375] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a227d83d-4ea4-4e96-a504-44c902590cb5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.953326] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fba3917-e40a-42e3-9e9f-69dbcc4daa23 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.961951] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5812520e-2801-4f95-a339-b2477dc80a7a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.977487] env[61243]: DEBUG nova.compute.provider_tree [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 558.200872] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Acquiring lock "84529438-8560-4323-a2f8-86cb917d18e6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.201610] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Lock "84529438-8560-4323-a2f8-86cb917d18e6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.240586] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': task-1338720, 'name': ReconfigVM_Task, 'duration_secs': 0.336781} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.240925] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 1b1b2bac-67af-4921-a49a-429017ef7a3b/1b1b2bac-67af-4921-a49a-429017ef7a3b.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 558.241646] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c1376e3c-e43c-4283-a278-d47bf897e5cb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.251170] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Waiting for the task: (returnval){ [ 558.251170] env[61243]: value = "task-1338721" [ 558.251170] env[61243]: _type = "Task" [ 558.251170] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.262035] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': task-1338721, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.373878] env[61243]: INFO nova.scheduler.client.report [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Deleted allocations for instance 84a32e46-ccf7-47ca-986c-8c59d5bf43a5 [ 558.505646] env[61243]: ERROR nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [req-ada4ab85-11c9-463e-a33b-4bf7511232ec] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ada4ab85-11c9-463e-a33b-4bf7511232ec"}]} [ 558.545075] env[61243]: DEBUG nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 558.564685] env[61243]: DEBUG nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 558.564920] env[61243]: DEBUG nova.compute.provider_tree [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 558.590841] env[61243]: DEBUG nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 558.620442] env[61243]: DEBUG nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 558.765989] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': task-1338721, 'name': Rename_Task, 'duration_secs': 0.151818} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.769340] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 558.769659] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a223454-b7d7-4a84-a0d3-a8280de258e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.781175] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Waiting for the task: (returnval){ [ 558.781175] env[61243]: value = "task-1338722" [ 558.781175] env[61243]: _type = "Task" [ 558.781175] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.791252] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': task-1338722, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.888406] env[61243]: DEBUG oslo_concurrency.lockutils [None req-39d920e4-1b34-4faa-b9fa-211d6003a98c tempest-ServersAdminNegativeTestJSON-36638245 tempest-ServersAdminNegativeTestJSON-36638245-project-member] Lock "84a32e46-ccf7-47ca-986c-8c59d5bf43a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.699s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.045485] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-765bb365-144d-4df6-b43d-3dd89501cf8a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.055229] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25eea742-18e0-46a9-ad59-eba25e8baa53 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.095788] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c611dece-a955-4fa0-8a2b-e013b3aba773 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.106750] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af09bb09-e326-4dc6-b226-49ef3191cc8e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.125208] env[61243]: DEBUG nova.compute.provider_tree [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 559.297521] env[61243]: DEBUG oslo_vmware.api [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Task: {'id': task-1338722, 'name': PowerOnVM_Task, 'duration_secs': 0.498247} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.297521] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 559.297521] env[61243]: DEBUG nova.compute.manager [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 559.298232] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db6f446-e8a8-4e91-855b-c3f875c6d2e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.390900] env[61243]: DEBUG nova.compute.manager [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 559.651773] env[61243]: ERROR nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [req-6dbf99fc-db93-4660-847f-b80f541d2aab] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6dbf99fc-db93-4660-847f-b80f541d2aab"}]} [ 559.681996] env[61243]: DEBUG nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 559.697787] env[61243]: DEBUG nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 559.698012] env[61243]: DEBUG nova.compute.provider_tree [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 559.711771] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Acquiring lock "4072d5a6-f69d-4fb9-837b-7855054e5206" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.711999] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Lock "4072d5a6-f69d-4fb9-837b-7855054e5206" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.724516] env[61243]: DEBUG nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 559.745842] env[61243]: DEBUG nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 559.821319] env[61243]: DEBUG oslo_concurrency.lockutils [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.917897] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.002021] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Acquiring lock "d113d9ef-5698-4851-a51a-7c7c807e9135" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.002021] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Lock "d113d9ef-5698-4851-a51a-7c7c807e9135" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.202431] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb6b08ca-38b1-4260-8fcb-ebefe23fafad {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.211162] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd112b5-de68-4f1c-a63a-8b053c61c604 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.248846] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67fb3e6-d154-4b55-9e98-db7e67895643 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.258532] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd3e43c-eec3-44ad-88cf-9779fd261431 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.275298] env[61243]: DEBUG nova.compute.provider_tree [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 560.799829] env[61243]: ERROR nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [req-1a20c17f-9810-4bde-9066-1d5012277e87] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1a20c17f-9810-4bde-9066-1d5012277e87"}]} [ 560.799829] env[61243]: DEBUG oslo_concurrency.lockutils [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 5.642s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.800692] env[61243]: ERROR nova.compute.manager [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] Failed to build and run instance: nova.exception.ResourceProviderUpdateConflict: A conflict was encountered attempting to update resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 (generation 36): {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1a20c17f-9810-4bde-9066-1d5012277e87"}]} [ 560.800692] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] Traceback (most recent call last): [ 560.800692] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] File "/opt/stack/nova/nova/compute/manager.py", line 2615, in _build_and_run_instance [ 560.800692] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] with self.rt.instance_claim(context, instance, node, allocs, [ 560.800692] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 560.800692] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] return f(*args, **kwargs) [ 560.800692] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 560.800692] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] self._update(elevated, cn) [ 560.800692] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 560.802977] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] self._update_to_placement(context, compute_node, startup) [ 560.802977] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 560.802977] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 560.802977] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 266, in call [ 560.802977] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] raise attempt.get() [ 560.802977] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 560.802977] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] six.reraise(self.value[0], self.value[1], self.value[2]) [ 560.802977] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 560.802977] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] raise value [ 560.802977] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 560.802977] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 560.802977] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 560.802977] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] self.reportclient.update_from_provider_tree( [ 560.803468] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 560.803468] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] self.set_inventory_for_provider( [ 560.803468] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1003, in set_inventory_for_provider [ 560.803468] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] raise exception.ResourceProviderUpdateConflict( [ 560.803468] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] nova.exception.ResourceProviderUpdateConflict: A conflict was encountered attempting to update resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 (generation 36): {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1a20c17f-9810-4bde-9066-1d5012277e87"}]} [ 560.803468] env[61243]: ERROR nova.compute.manager [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] [ 560.803468] env[61243]: DEBUG nova.compute.utils [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] A conflict was encountered attempting to update resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 (generation 36): {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource pro {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 560.803725] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.889s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.809616] env[61243]: DEBUG nova.compute.manager [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] Build of instance e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044 was re-scheduled: A conflict was encountered attempting to update resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 (generation 36): {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1a20c17f-9810-4bde-9066-1d5012277e87"}]} {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 560.810090] env[61243]: DEBUG nova.compute.manager [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 560.810267] env[61243]: DEBUG oslo_concurrency.lockutils [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Acquiring lock "refresh_cache-e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.810412] env[61243]: DEBUG oslo_concurrency.lockutils [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Acquired lock "refresh_cache-e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.810594] env[61243]: DEBUG nova.network.neutron [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 561.019798] env[61243]: DEBUG oslo_concurrency.lockutils [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquiring lock "1b1b2bac-67af-4921-a49a-429017ef7a3b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.020847] env[61243]: DEBUG oslo_concurrency.lockutils [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Lock "1b1b2bac-67af-4921-a49a-429017ef7a3b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.020847] env[61243]: DEBUG oslo_concurrency.lockutils [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquiring lock "1b1b2bac-67af-4921-a49a-429017ef7a3b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.020847] env[61243]: DEBUG oslo_concurrency.lockutils [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Lock "1b1b2bac-67af-4921-a49a-429017ef7a3b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.020847] env[61243]: DEBUG oslo_concurrency.lockutils [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Lock "1b1b2bac-67af-4921-a49a-429017ef7a3b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.027713] env[61243]: INFO nova.compute.manager [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Terminating instance [ 561.028367] env[61243]: DEBUG oslo_concurrency.lockutils [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquiring lock "refresh_cache-1b1b2bac-67af-4921-a49a-429017ef7a3b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.028367] env[61243]: DEBUG oslo_concurrency.lockutils [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquired lock "refresh_cache-1b1b2bac-67af-4921-a49a-429017ef7a3b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.028465] env[61243]: DEBUG nova.network.neutron [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 561.052721] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Acquiring lock "06264429-11eb-4d64-9f6b-6587cf048c41" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.052978] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Lock "06264429-11eb-4d64-9f6b-6587cf048c41" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.338408] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 561.348833] env[61243]: DEBUG nova.network.neutron [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.357526] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 561.357851] env[61243]: DEBUG nova.compute.provider_tree [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 561.378478] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 561.401073] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 561.550612] env[61243]: DEBUG nova.network.neutron [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.566209] env[61243]: DEBUG nova.network.neutron [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.650915] env[61243]: DEBUG nova.network.neutron [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.892631] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd7226bd-bb28-4b85-aa67-7d148cd50720 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.901370] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b45db54-acd8-493f-9729-23167e84777a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.935427] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6137d9-d0c7-4452-9ac8-7b420ecc030f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.945113] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c06d78da-d4de-4ea4-927f-63a5b5143b9b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.963279] env[61243]: DEBUG nova.compute.provider_tree [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 562.053110] env[61243]: DEBUG oslo_concurrency.lockutils [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Releasing lock "refresh_cache-e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.053110] env[61243]: DEBUG nova.compute.manager [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 562.053110] env[61243]: DEBUG nova.compute.manager [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 562.053333] env[61243]: DEBUG nova.network.neutron [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 562.077678] env[61243]: DEBUG nova.network.neutron [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.155122] env[61243]: DEBUG oslo_concurrency.lockutils [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Releasing lock "refresh_cache-1b1b2bac-67af-4921-a49a-429017ef7a3b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.155122] env[61243]: DEBUG nova.compute.manager [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 562.155339] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 562.156781] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fab0a6f-d2b9-4a7f-ada2-a7b95de93dfc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.167659] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 562.168106] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d523e0a1-b6dc-4bca-a0e7-10e13ea0eda3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.177718] env[61243]: DEBUG oslo_vmware.api [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 562.177718] env[61243]: value = "task-1338723" [ 562.177718] env[61243]: _type = "Task" [ 562.177718] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.189770] env[61243]: DEBUG oslo_vmware.api [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338723, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.490635] env[61243]: ERROR nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [req-f73065a4-5f65-4cab-85aa-f3b93b788236] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f73065a4-5f65-4cab-85aa-f3b93b788236"}]}: nova.exception.PortBindingFailed: Binding failed for port 937400db-3ff6-4e49-971b-479c6d6b0fc8, please check neutron logs for more information. [ 562.512530] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 562.527767] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 562.527767] env[61243]: DEBUG nova.compute.provider_tree [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 562.542341] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: 363c7a25-4987-4af8-b63d-f05839caabb4 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 562.561165] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 562.582574] env[61243]: DEBUG nova.network.neutron [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.694748] env[61243]: DEBUG oslo_vmware.api [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338723, 'name': PowerOffVM_Task, 'duration_secs': 0.131946} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.695648] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 562.695648] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 562.695818] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5446795e-86f9-406b-94d4-680647e9ee31 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.743774] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 562.743774] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 562.743774] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Deleting the datastore file [datastore2] 1b1b2bac-67af-4921-a49a-429017ef7a3b {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 562.743774] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a60f9a1c-93cf-4a67-a1fc-a7a8ed4ce2e7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.756537] env[61243]: DEBUG oslo_vmware.api [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for the task: (returnval){ [ 562.756537] env[61243]: value = "task-1338725" [ 562.756537] env[61243]: _type = "Task" [ 562.756537] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.770652] env[61243]: DEBUG oslo_vmware.api [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338725, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.067735] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "d5131a9f-f4bd-40c8-98d1-043651537a16" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.068039] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "d5131a9f-f4bd-40c8-98d1-043651537a16" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.090018] env[61243]: INFO nova.compute.manager [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] [instance: e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044] Took 1.04 seconds to deallocate network for instance. [ 563.138269] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea8900f7-1bf0-4555-8dcc-9614b943038c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.153147] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9609b8-08da-4a22-85d8-5df44fc7ac9c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.186975] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0971c8-47fe-4625-81dc-35195c082141 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.196741] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb65a59-521c-4dcd-bff4-c4b86db5e1d7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.213957] env[61243]: DEBUG nova.compute.provider_tree [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 563.269166] env[61243]: DEBUG oslo_vmware.api [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Task: {'id': task-1338725, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.097303} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.269453] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 563.269677] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 563.269964] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 563.270080] env[61243]: INFO nova.compute.manager [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 563.270374] env[61243]: DEBUG oslo.service.loopingcall [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 563.270652] env[61243]: DEBUG nova.compute.manager [-] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 563.270751] env[61243]: DEBUG nova.network.neutron [-] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 563.300138] env[61243]: DEBUG nova.network.neutron [-] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.750353] env[61243]: ERROR nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [req-2c3af1da-ef1a-4988-9998-69e9f2c75838] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2c3af1da-ef1a-4988-9998-69e9f2c75838"}]}: nova.exception.PortBindingFailed: Binding failed for port 937400db-3ff6-4e49-971b-479c6d6b0fc8, please check neutron logs for more information. [ 563.775186] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 563.801504] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 563.801731] env[61243]: DEBUG nova.compute.provider_tree [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 563.803822] env[61243]: DEBUG nova.network.neutron [-] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.819567] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 563.866054] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 564.177469] env[61243]: INFO nova.scheduler.client.report [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Deleted allocations for instance e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044 [ 564.316789] env[61243]: INFO nova.compute.manager [-] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Took 1.04 seconds to deallocate network for instance. [ 564.466155] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d4fdf10-4adc-4c75-808a-5ed76c92e48a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.475333] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f783def5-c0dd-4e3b-8902-18f9527e84f3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.510110] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5173a76f-2c18-40b8-867b-32ec8db00e9d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.519033] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f809860d-d88d-4488-9f9c-0fcc0cfbf7cb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.534881] env[61243]: DEBUG nova.compute.provider_tree [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 564.693340] env[61243]: DEBUG oslo_concurrency.lockutils [None req-474ca366-ca4b-4ba5-8fe0-78e83ed4448b tempest-VolumesAssistedSnapshotsTest-961157475 tempest-VolumesAssistedSnapshotsTest-961157475-project-member] Lock "e0ae2a1d-5cbf-4d8a-8a3e-aac86bff5044" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.429s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.831397] env[61243]: DEBUG oslo_concurrency.lockutils [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.062220] env[61243]: ERROR nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [req-358c1478-533c-4031-b768-53032aa27ed0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-358c1478-533c-4031-b768-53032aa27ed0"}]}: nova.exception.PortBindingFailed: Binding failed for port 937400db-3ff6-4e49-971b-479c6d6b0fc8, please check neutron logs for more information. [ 565.081699] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 565.106254] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 565.106254] env[61243]: DEBUG nova.compute.provider_tree [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 565.134110] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 565.157647] env[61243]: DEBUG nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 565.195920] env[61243]: DEBUG nova.compute.manager [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 565.388632] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Acquiring lock "72aa2816-5e02-4dc2-9d6a-c3e58c651224" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.388850] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Lock "72aa2816-5e02-4dc2-9d6a-c3e58c651224" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.589038] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b03d3a-3d14-4b21-bdb1-eee64833d272 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.598321] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-522cac2a-4efb-40a1-acda-f600df472ad3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.634214] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce36d20-82f6-4a28-a148-f80de9b8c874 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.641570] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b09f542-70c7-4162-a904-bde510d29834 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.655349] env[61243]: DEBUG nova.compute.provider_tree [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 565.719301] env[61243]: DEBUG oslo_concurrency.lockutils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.182870] env[61243]: ERROR nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [req-c9869f24-2b5b-4b56-b64f-c72408caa23a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c9869f24-2b5b-4b56-b64f-c72408caa23a"}]}: nova.exception.PortBindingFailed: Binding failed for port 937400db-3ff6-4e49-971b-479c6d6b0fc8, please check neutron logs for more information. [ 566.182870] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 5.378s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.183278] env[61243]: ERROR nova.compute.manager [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Failed to build and run instance: nova.exception.ResourceProviderUpdateConflict: A conflict was encountered attempting to update resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 (generation 41): {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c9869f24-2b5b-4b56-b64f-c72408caa23a"}]} [ 566.183278] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Traceback (most recent call last): [ 566.183278] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 566.183278] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] self.driver.spawn(context, instance, image_meta, [ 566.183278] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 566.183278] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] self._vmops.spawn(context, instance, image_meta, injected_files, [ 566.183278] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 566.183278] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] vm_ref = self.build_virtual_machine(instance, [ 566.183278] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 566.183575] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] vif_infos = vmwarevif.get_vif_info(self._session, [ 566.183575] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 566.183575] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] for vif in network_info: [ 566.183575] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 566.183575] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] return self._sync_wrapper(fn, *args, **kwargs) [ 566.183575] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 566.183575] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] self.wait() [ 566.183575] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 566.183575] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] self[:] = self._gt.wait() [ 566.183575] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 566.183575] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] return self._exit_event.wait() [ 566.183575] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 566.183575] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] result = hub.switch() [ 566.183922] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 566.183922] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] return self.greenlet.switch() [ 566.183922] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 566.183922] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] result = function(*args, **kwargs) [ 566.183922] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 566.183922] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] return func(*args, **kwargs) [ 566.183922] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 566.183922] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] raise e [ 566.183922] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.183922] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] nwinfo = self.network_api.allocate_for_instance( [ 566.183922] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 566.183922] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] created_port_ids = self._update_ports_for_instance( [ 566.183922] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 566.184431] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] with excutils.save_and_reraise_exception(): [ 566.184431] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.184431] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] self.force_reraise() [ 566.184431] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.184431] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] raise self.value [ 566.184431] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 566.184431] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] updated_port = self._update_port( [ 566.184431] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.184431] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] _ensure_no_port_binding_failure(port) [ 566.184431] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.184431] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] raise exception.PortBindingFailed(port_id=port['id']) [ 566.184431] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] nova.exception.PortBindingFailed: Binding failed for port 937400db-3ff6-4e49-971b-479c6d6b0fc8, please check neutron logs for more information. [ 566.184431] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] [ 566.184776] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] During handling of the above exception, another exception occurred: [ 566.184776] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] [ 566.184776] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Traceback (most recent call last): [ 566.184776] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/compute/manager.py", line 2615, in _build_and_run_instance [ 566.184776] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] with self.rt.instance_claim(context, instance, node, allocs, [ 566.184776] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/compute/claims.py", line 43, in __exit__ [ 566.184776] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] self.abort() [ 566.184776] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/compute/claims.py", line 86, in abort [ 566.184776] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] self.tracker.abort_instance_claim(self.context, self.instance_ref, [ 566.184776] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 566.184776] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] return f(*args, **kwargs) [ 566.184776] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 586, in abort_instance_claim [ 566.184776] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] self._update(context.elevated(), self.compute_nodes[nodename]) [ 566.185200] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 566.185200] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] self._update_to_placement(context, compute_node, startup) [ 566.185200] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 566.185200] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 566.185200] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 266, in call [ 566.185200] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] raise attempt.get() [ 566.185200] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 566.185200] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] six.reraise(self.value[0], self.value[1], self.value[2]) [ 566.185200] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 566.185200] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] raise value [ 566.185200] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 566.185200] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 566.186088] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 566.186088] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] self.reportclient.update_from_provider_tree( [ 566.186088] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 566.186088] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] self.set_inventory_for_provider( [ 566.186088] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1003, in set_inventory_for_provider [ 566.186088] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] raise exception.ResourceProviderUpdateConflict( [ 566.186088] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] nova.exception.ResourceProviderUpdateConflict: A conflict was encountered attempting to update resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 (generation 41): {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c9869f24-2b5b-4b56-b64f-c72408caa23a"}]} [ 566.186088] env[61243]: ERROR nova.compute.manager [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] [ 566.186501] env[61243]: DEBUG nova.compute.utils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] A conflict was encountered attempting to update resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 (generation 41): {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource pro {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 566.186501] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 26.010s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.188146] env[61243]: DEBUG nova.compute.manager [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Build of instance 6c4e205a-6f9d-4439-add7-376fe718d189 was re-scheduled: A conflict was encountered attempting to update resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 (generation 41): {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c9869f24-2b5b-4b56-b64f-c72408caa23a"}]} {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 566.188738] env[61243]: DEBUG nova.compute.manager [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 566.188961] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Acquiring lock "refresh_cache-6c4e205a-6f9d-4439-add7-376fe718d189" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.189121] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Acquired lock "refresh_cache-6c4e205a-6f9d-4439-add7-376fe718d189" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.189279] env[61243]: DEBUG nova.network.neutron [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 566.547355] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Acquiring lock "c887c01b-a67e-4e3e-9e17-1deb26e9b508" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.547668] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Lock "c887c01b-a67e-4e3e-9e17-1deb26e9b508" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.715842] env[61243]: DEBUG nova.network.neutron [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.808203] env[61243]: DEBUG nova.network.neutron [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.243147] env[61243]: WARNING nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 1b1b2bac-67af-4921-a49a-429017ef7a3b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 567.312012] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Releasing lock "refresh_cache-6c4e205a-6f9d-4439-add7-376fe718d189" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.312012] env[61243]: DEBUG nova.compute.manager [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 567.312423] env[61243]: DEBUG nova.compute.manager [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 567.312497] env[61243]: DEBUG nova.network.neutron [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 567.338908] env[61243]: DEBUG nova.network.neutron [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.748579] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 6c4e205a-6f9d-4439-add7-376fe718d189 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 567.749173] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance da977758-95d5-4fb6-aaa3-7b832404b527 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 567.749444] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 567.749675] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 25a896cb-fd91-4f16-a8cd-6de58fcfcc27 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 567.750057] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 8c93ef89-5d27-48d3-8c44-79eb23d66ee3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 567.845418] env[61243]: DEBUG nova.network.neutron [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.253716] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 568.349450] env[61243]: INFO nova.compute.manager [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 6c4e205a-6f9d-4439-add7-376fe718d189] Took 1.04 seconds to deallocate network for instance. [ 568.761102] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 762cbca7-4729-458a-b2af-2978723a25dc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 568.907566] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Acquiring lock "cceaed50-e20f-45e4-9959-29dac2ecad30" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.909735] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Lock "cceaed50-e20f-45e4-9959-29dac2ecad30" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.265500] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 3700deef-a438-4920-b35f-8fba370a5984 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 569.399463] env[61243]: INFO nova.scheduler.client.report [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Deleted allocations for instance 6c4e205a-6f9d-4439-add7-376fe718d189 [ 569.772495] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 2451843a-fd32-450e-933e-e635c789834b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 569.916370] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ec5c6c93-8a55-40c8-a0ea-aa81a4b52306 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Lock "6c4e205a-6f9d-4439-add7-376fe718d189" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.164s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.275880] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance f21b7977-747e-437f-b03d-8442ed58756e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 570.422769] env[61243]: DEBUG nova.compute.manager [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 570.781512] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance e2c498c5-3df1-4b99-a322-a598afd506f8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 570.954975] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.285926] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 5508a717-fb58-49b3-bfd1-b80d76a3c4e1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 571.790364] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 476632bb-e07c-4660-9ea4-7290884e4761 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 572.296994] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance b5163d89-ceb8-4c61-ae02-0ae3311b58a9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 572.800311] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 461b2ef2-d665-47ab-affa-e525bffe561a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 573.304164] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance b3d98e6f-4b45-4eaf-a79c-b683b32fe053 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 573.377261] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Acquiring lock "755cdc16-0eae-4978-917f-634a8ac7af8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.377600] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Lock "755cdc16-0eae-4978-917f-634a8ac7af8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.809093] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 0b223791-7e95-43fe-bef0-bac6222884ea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 574.312596] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 64efd676-9166-4ffa-a437-6fa6c466ace0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 574.816643] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 0625ac10-55b4-488a-aba5-bdaa394045f4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 575.088910] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Acquiring lock "37b43ee9-f991-4d56-8fcc-192f6534f803" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.089386] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Lock "37b43ee9-f991-4d56-8fcc-192f6534f803" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.320271] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 8efd391b-7b76-4167-bd18-48566271ccc8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 575.823363] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 84529438-8560-4323-a2f8-86cb917d18e6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 576.326864] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 4072d5a6-f69d-4fb9-837b-7855054e5206 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 576.831934] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance d113d9ef-5698-4851-a51a-7c7c807e9135 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 577.065203] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Acquiring lock "8e85f285-1190-4c15-bc1a-8364434dc4ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.065416] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Lock "8e85f285-1190-4c15-bc1a-8364434dc4ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.337034] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 06264429-11eb-4d64-9f6b-6587cf048c41 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 577.844438] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance d5131a9f-f4bd-40c8-98d1-043651537a16 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 578.092292] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Acquiring lock "f6566092-9921-4a94-a0b5-a2d3803a9d18" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.092419] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Lock "f6566092-9921-4a94-a0b5-a2d3803a9d18" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.345924] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 72aa2816-5e02-4dc2-9d6a-c3e58c651224 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 578.851525] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance c887c01b-a67e-4e3e-9e17-1deb26e9b508 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 578.851848] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 578.851934] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 578.873031] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 578.889993] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 578.890239] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 578.902057] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 578.919902] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 579.426440] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40acf655-32e8-420a-b1a0-a4f35fa07a66 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.435011] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a550f893-73aa-4494-9e4a-b5427ef6c6dc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.466193] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f33ecc48-82c1-4f36-b890-baff7c5a7058 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.473764] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2ce50a-5a87-4539-a536-cad57e4ce21d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.486768] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 579.980060] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Acquiring lock "fd2d724a-76b6-405f-b003-54de2ad84f5a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.980879] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Lock "fd2d724a-76b6-405f-b003-54de2ad84f5a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.010029] env[61243]: ERROR nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [req-60b3b012-3385-4f42-a07e-9bc070ee2573] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-60b3b012-3385-4f42-a07e-9bc070ee2573"}]} [ 580.028986] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 580.047521] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 580.047761] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 580.062133] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 580.086587] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 580.497646] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca708e0-02c3-4ca0-a1d3-26a10b2c5ea1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.507461] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca61d81c-c69a-4f59-8352-2fae5c364557 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.536914] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd2c5e15-87d6-4e9c-9365-5563bc492684 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.545501] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48cef0fa-6251-4aa6-b39a-0ba1c44f391b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.558985] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 581.112447] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 48 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 581.112736] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 48 to 49 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 581.112825] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 581.618354] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61243) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 581.618964] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 15.435s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.619272] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 39.915s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.637285] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4303d23-4c06-49d9-83f6-409874b6dbe6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.645512] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2661781-e0b6-4102-85c6-2f3fc48b0456 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.678695] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643debf2-3268-4bb7-9ead-4e2bbec27a39 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.686810] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35dc26ab-ee78-4dd4-88f8-ac74f416b994 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.704470] env[61243]: DEBUG nova.compute.provider_tree [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.208739] env[61243]: DEBUG nova.scheduler.client.report [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 583.714949] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.096s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.715656] env[61243]: ERROR nova.compute.manager [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5847ee11-06ae-4d8d-bba0-419f07abff41, please check neutron logs for more information. [ 583.715656] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Traceback (most recent call last): [ 583.715656] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 583.715656] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] self.driver.spawn(context, instance, image_meta, [ 583.715656] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 583.715656] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] self._vmops.spawn(context, instance, image_meta, injected_files, [ 583.715656] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 583.715656] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] vm_ref = self.build_virtual_machine(instance, [ 583.715656] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 583.715656] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] vif_infos = vmwarevif.get_vif_info(self._session, [ 583.715656] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 583.716082] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] for vif in network_info: [ 583.716082] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 583.716082] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] return self._sync_wrapper(fn, *args, **kwargs) [ 583.716082] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 583.716082] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] self.wait() [ 583.716082] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 583.716082] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] self[:] = self._gt.wait() [ 583.716082] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 583.716082] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] return self._exit_event.wait() [ 583.716082] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 583.716082] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] result = hub.switch() [ 583.716082] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 583.716082] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] return self.greenlet.switch() [ 583.716763] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.716763] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] result = function(*args, **kwargs) [ 583.716763] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 583.716763] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] return func(*args, **kwargs) [ 583.716763] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.716763] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] raise e [ 583.716763] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.716763] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] nwinfo = self.network_api.allocate_for_instance( [ 583.716763] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.716763] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] created_port_ids = self._update_ports_for_instance( [ 583.716763] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.716763] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] with excutils.save_and_reraise_exception(): [ 583.716763] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.718148] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] self.force_reraise() [ 583.718148] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.718148] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] raise self.value [ 583.718148] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.718148] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] updated_port = self._update_port( [ 583.718148] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.718148] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] _ensure_no_port_binding_failure(port) [ 583.718148] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.718148] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] raise exception.PortBindingFailed(port_id=port['id']) [ 583.718148] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] nova.exception.PortBindingFailed: Binding failed for port 5847ee11-06ae-4d8d-bba0-419f07abff41, please check neutron logs for more information. [ 583.718148] env[61243]: ERROR nova.compute.manager [instance: da977758-95d5-4fb6-aaa3-7b832404b527] [ 583.718458] env[61243]: DEBUG nova.compute.utils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Binding failed for port 5847ee11-06ae-4d8d-bba0-419f07abff41, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 583.718458] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 41.105s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.719490] env[61243]: INFO nova.compute.claims [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 583.722650] env[61243]: DEBUG nova.compute.manager [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Build of instance da977758-95d5-4fb6-aaa3-7b832404b527 was re-scheduled: Binding failed for port 5847ee11-06ae-4d8d-bba0-419f07abff41, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 583.723172] env[61243]: DEBUG nova.compute.manager [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 583.723345] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Acquiring lock "refresh_cache-da977758-95d5-4fb6-aaa3-7b832404b527" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.723486] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Acquired lock "refresh_cache-da977758-95d5-4fb6-aaa3-7b832404b527" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.723640] env[61243]: DEBUG nova.network.neutron [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 583.905955] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Acquiring lock "3b852a19-9f86-486b-b98d-e87ffc54d943" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.906369] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Lock "3b852a19-9f86-486b-b98d-e87ffc54d943" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.248237] env[61243]: DEBUG nova.network.neutron [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.300937] env[61243]: DEBUG nova.network.neutron [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.805888] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Releasing lock "refresh_cache-da977758-95d5-4fb6-aaa3-7b832404b527" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.806374] env[61243]: DEBUG nova.compute.manager [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 584.806771] env[61243]: DEBUG nova.compute.manager [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 584.807047] env[61243]: DEBUG nova.network.neutron [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 584.830948] env[61243]: DEBUG nova.network.neutron [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.272031] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ec3126-f75c-4f31-b2cf-7813a8c78fbb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.280022] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f8c171-83b0-4bb4-b424-578d6196a565 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.311884] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0531715a-0058-4490-a893-b1c727385e30 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.320120] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b93a63-5cca-4b6c-ba63-f3b3c03ca090 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.334367] env[61243]: DEBUG nova.compute.provider_tree [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.336169] env[61243]: DEBUG nova.network.neutron [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.839113] env[61243]: DEBUG nova.scheduler.client.report [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 585.846252] env[61243]: INFO nova.compute.manager [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: da977758-95d5-4fb6-aaa3-7b832404b527] Took 1.04 seconds to deallocate network for instance. [ 586.347812] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.630s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.348163] env[61243]: DEBUG nova.compute.manager [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 586.350718] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 43.585s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.863261] env[61243]: DEBUG nova.compute.utils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 586.867560] env[61243]: DEBUG nova.compute.manager [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 586.867723] env[61243]: DEBUG nova.network.neutron [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 586.908783] env[61243]: INFO nova.scheduler.client.report [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Deleted allocations for instance da977758-95d5-4fb6-aaa3-7b832404b527 [ 586.926192] env[61243]: DEBUG nova.policy [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a9e6dc85a31d4ec38efe0c750c49ead1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '113713170fb94b8a8d7985b9a02faf77', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 587.145308] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Acquiring lock "be8ac265-b214-4db5-9781-d171cee46538" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.145603] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Lock "be8ac265-b214-4db5-9781-d171cee46538" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.368528] env[61243]: DEBUG nova.compute.manager [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 587.379314] env[61243]: DEBUG nova.network.neutron [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Successfully created port: 0ebbe785-0120-4fff-a6a9-8a38588de2ee {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 587.419026] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5a5c083e-a206-47ca-9037-972c2343e538 tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Lock "da977758-95d5-4fb6-aaa3-7b832404b527" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.047s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.424318] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ca3e23-a6d6-4942-8bbe-db162df6e2ad {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.433748] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5ad7145-88cd-4bfc-867e-2ce82e24311f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.466270] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0272f1da-2a8b-484e-8fa3-4704d974a5b1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.475547] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e13e6358-990a-4e5b-bf7b-163e5575a6d5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.490862] env[61243]: DEBUG nova.compute.provider_tree [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.923997] env[61243]: DEBUG nova.compute.manager [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 587.997201] env[61243]: DEBUG nova.scheduler.client.report [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 588.335602] env[61243]: ERROR nova.compute.manager [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0ebbe785-0120-4fff-a6a9-8a38588de2ee, please check neutron logs for more information. [ 588.335602] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 588.335602] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.335602] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 588.335602] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.335602] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 588.335602] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.335602] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 588.335602] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.335602] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 588.335602] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.335602] env[61243]: ERROR nova.compute.manager raise self.value [ 588.335602] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.335602] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 588.335602] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.335602] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 588.336183] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.336183] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 588.336183] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0ebbe785-0120-4fff-a6a9-8a38588de2ee, please check neutron logs for more information. [ 588.336183] env[61243]: ERROR nova.compute.manager [ 588.336183] env[61243]: Traceback (most recent call last): [ 588.336183] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 588.336183] env[61243]: listener.cb(fileno) [ 588.336183] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.336183] env[61243]: result = function(*args, **kwargs) [ 588.336183] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 588.336183] env[61243]: return func(*args, **kwargs) [ 588.336183] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.336183] env[61243]: raise e [ 588.336183] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.336183] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 588.336183] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.336183] env[61243]: created_port_ids = self._update_ports_for_instance( [ 588.336183] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.336183] env[61243]: with excutils.save_and_reraise_exception(): [ 588.336183] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.336183] env[61243]: self.force_reraise() [ 588.336183] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.336183] env[61243]: raise self.value [ 588.336183] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.336183] env[61243]: updated_port = self._update_port( [ 588.336183] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.336183] env[61243]: _ensure_no_port_binding_failure(port) [ 588.336183] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.336183] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 588.337632] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 0ebbe785-0120-4fff-a6a9-8a38588de2ee, please check neutron logs for more information. [ 588.337632] env[61243]: Removing descriptor: 15 [ 588.337632] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Acquiring lock "e02f41be-125a-4228-8136-9e6639a268a9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.337632] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Lock "e02f41be-125a-4228-8136-9e6639a268a9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.378949] env[61243]: DEBUG nova.compute.manager [req-e5faf59d-db85-481c-a1cd-a0cfa3cb9c25 req-7ee48829-afc3-46d1-841f-4a1b0f2dd57b service nova] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Received event network-changed-0ebbe785-0120-4fff-a6a9-8a38588de2ee {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 588.379171] env[61243]: DEBUG nova.compute.manager [req-e5faf59d-db85-481c-a1cd-a0cfa3cb9c25 req-7ee48829-afc3-46d1-841f-4a1b0f2dd57b service nova] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Refreshing instance network info cache due to event network-changed-0ebbe785-0120-4fff-a6a9-8a38588de2ee. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 588.379383] env[61243]: DEBUG oslo_concurrency.lockutils [req-e5faf59d-db85-481c-a1cd-a0cfa3cb9c25 req-7ee48829-afc3-46d1-841f-4a1b0f2dd57b service nova] Acquiring lock "refresh_cache-9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.380630] env[61243]: DEBUG oslo_concurrency.lockutils [req-e5faf59d-db85-481c-a1cd-a0cfa3cb9c25 req-7ee48829-afc3-46d1-841f-4a1b0f2dd57b service nova] Acquired lock "refresh_cache-9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.380630] env[61243]: DEBUG nova.network.neutron [req-e5faf59d-db85-481c-a1cd-a0cfa3cb9c25 req-7ee48829-afc3-46d1-841f-4a1b0f2dd57b service nova] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Refreshing network info cache for port 0ebbe785-0120-4fff-a6a9-8a38588de2ee {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 588.388103] env[61243]: DEBUG nova.compute.manager [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 588.420451] env[61243]: DEBUG nova.virt.hardware [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 588.420753] env[61243]: DEBUG nova.virt.hardware [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 588.422100] env[61243]: DEBUG nova.virt.hardware [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 588.422100] env[61243]: DEBUG nova.virt.hardware [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 588.422100] env[61243]: DEBUG nova.virt.hardware [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 588.422100] env[61243]: DEBUG nova.virt.hardware [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 588.422100] env[61243]: DEBUG nova.virt.hardware [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 588.422463] env[61243]: DEBUG nova.virt.hardware [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 588.423864] env[61243]: DEBUG nova.virt.hardware [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 588.424068] env[61243]: DEBUG nova.virt.hardware [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 588.424249] env[61243]: DEBUG nova.virt.hardware [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 588.425877] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d6e5084-aa33-4f83-86ef-f85e88978436 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.436587] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5c27c3-6773-4b0e-b26b-ab5ffd57903d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.455407] env[61243]: ERROR nova.compute.manager [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0ebbe785-0120-4fff-a6a9-8a38588de2ee, please check neutron logs for more information. [ 588.455407] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Traceback (most recent call last): [ 588.455407] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 588.455407] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] yield resources [ 588.455407] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.455407] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] self.driver.spawn(context, instance, image_meta, [ 588.455407] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 588.455407] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.455407] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.455407] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] vm_ref = self.build_virtual_machine(instance, [ 588.455407] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.455854] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.455854] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.455854] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] for vif in network_info: [ 588.455854] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.455854] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] return self._sync_wrapper(fn, *args, **kwargs) [ 588.455854] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.455854] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] self.wait() [ 588.455854] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.455854] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] self[:] = self._gt.wait() [ 588.455854] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.455854] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] return self._exit_event.wait() [ 588.455854] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 588.455854] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] current.throw(*self._exc) [ 588.456279] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.456279] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] result = function(*args, **kwargs) [ 588.456279] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 588.456279] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] return func(*args, **kwargs) [ 588.456279] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.456279] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] raise e [ 588.456279] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.456279] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] nwinfo = self.network_api.allocate_for_instance( [ 588.456279] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.456279] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] created_port_ids = self._update_ports_for_instance( [ 588.456279] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.456279] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] with excutils.save_and_reraise_exception(): [ 588.456279] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.456649] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] self.force_reraise() [ 588.456649] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.456649] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] raise self.value [ 588.456649] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.456649] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] updated_port = self._update_port( [ 588.456649] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.456649] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] _ensure_no_port_binding_failure(port) [ 588.456649] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.456649] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] raise exception.PortBindingFailed(port_id=port['id']) [ 588.456649] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] nova.exception.PortBindingFailed: Binding failed for port 0ebbe785-0120-4fff-a6a9-8a38588de2ee, please check neutron logs for more information. [ 588.456649] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] [ 588.456649] env[61243]: INFO nova.compute.manager [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Terminating instance [ 588.458057] env[61243]: DEBUG oslo_concurrency.lockutils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.458801] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "refresh_cache-9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.500308] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.149s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.502340] env[61243]: ERROR nova.compute.manager [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d044cd3a-6023-4270-aa0d-08b4b677f8f3, please check neutron logs for more information. [ 588.502340] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Traceback (most recent call last): [ 588.502340] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.502340] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] self.driver.spawn(context, instance, image_meta, [ 588.502340] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 588.502340] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.502340] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.502340] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] vm_ref = self.build_virtual_machine(instance, [ 588.502340] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.502340] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.502340] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.502759] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] for vif in network_info: [ 588.502759] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.502759] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] return self._sync_wrapper(fn, *args, **kwargs) [ 588.502759] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.502759] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] self.wait() [ 588.502759] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.502759] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] self[:] = self._gt.wait() [ 588.502759] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.502759] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] return self._exit_event.wait() [ 588.502759] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 588.502759] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] result = hub.switch() [ 588.502759] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 588.502759] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] return self.greenlet.switch() [ 588.503215] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.503215] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] result = function(*args, **kwargs) [ 588.503215] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 588.503215] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] return func(*args, **kwargs) [ 588.503215] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.503215] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] raise e [ 588.503215] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.503215] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] nwinfo = self.network_api.allocate_for_instance( [ 588.503215] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.503215] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] created_port_ids = self._update_ports_for_instance( [ 588.503215] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.503215] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] with excutils.save_and_reraise_exception(): [ 588.503215] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.503635] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] self.force_reraise() [ 588.503635] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.503635] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] raise self.value [ 588.503635] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.503635] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] updated_port = self._update_port( [ 588.503635] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.503635] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] _ensure_no_port_binding_failure(port) [ 588.503635] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.503635] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] raise exception.PortBindingFailed(port_id=port['id']) [ 588.503635] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] nova.exception.PortBindingFailed: Binding failed for port d044cd3a-6023-4270-aa0d-08b4b677f8f3, please check neutron logs for more information. [ 588.503635] env[61243]: ERROR nova.compute.manager [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] [ 588.503996] env[61243]: DEBUG nova.compute.utils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Binding failed for port d044cd3a-6023-4270-aa0d-08b4b677f8f3, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 588.503996] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 40.056s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.503996] env[61243]: DEBUG nova.objects.instance [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61243) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 588.508034] env[61243]: DEBUG nova.compute.manager [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Build of instance f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d was re-scheduled: Binding failed for port d044cd3a-6023-4270-aa0d-08b4b677f8f3, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 588.508486] env[61243]: DEBUG nova.compute.manager [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 588.508710] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Acquiring lock "refresh_cache-f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.508855] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Acquired lock "refresh_cache-f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.509017] env[61243]: DEBUG nova.network.neutron [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.898708] env[61243]: DEBUG nova.network.neutron [req-e5faf59d-db85-481c-a1cd-a0cfa3cb9c25 req-7ee48829-afc3-46d1-841f-4a1b0f2dd57b service nova] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.985708] env[61243]: DEBUG nova.network.neutron [req-e5faf59d-db85-481c-a1cd-a0cfa3cb9c25 req-7ee48829-afc3-46d1-841f-4a1b0f2dd57b service nova] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.026282] env[61243]: DEBUG nova.network.neutron [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.094827] env[61243]: DEBUG nova.network.neutron [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.489445] env[61243]: DEBUG oslo_concurrency.lockutils [req-e5faf59d-db85-481c-a1cd-a0cfa3cb9c25 req-7ee48829-afc3-46d1-841f-4a1b0f2dd57b service nova] Releasing lock "refresh_cache-9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.489873] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "refresh_cache-9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.490333] env[61243]: DEBUG nova.network.neutron [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 589.514919] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50b2e238-7646-4d5e-bec8-21d699f76b7e tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.517367] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Acquiring lock "b1bad82d-d7bf-4e58-bbf2-f3d68f214115" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.517553] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Lock "b1bad82d-d7bf-4e58-bbf2-f3d68f214115" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.518156] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 40.841s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.519540] env[61243]: INFO nova.compute.claims [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 589.598140] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Releasing lock "refresh_cache-f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.598369] env[61243]: DEBUG nova.compute.manager [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 589.598594] env[61243]: DEBUG nova.compute.manager [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.598720] env[61243]: DEBUG nova.network.neutron [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 589.613810] env[61243]: DEBUG nova.network.neutron [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.016581] env[61243]: DEBUG nova.network.neutron [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.120250] env[61243]: DEBUG nova.network.neutron [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.131063] env[61243]: DEBUG nova.network.neutron [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.403539] env[61243]: DEBUG nova.compute.manager [req-649d3f57-e130-41de-89b0-c12b3f90a184 req-35bf6854-bb57-4ce6-ab31-cfab48a8e3a0 service nova] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Received event network-vif-deleted-0ebbe785-0120-4fff-a6a9-8a38588de2ee {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 590.624588] env[61243]: INFO nova.compute.manager [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] [instance: f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d] Took 1.03 seconds to deallocate network for instance. [ 590.635634] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "refresh_cache-9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.636042] env[61243]: DEBUG nova.compute.manager [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 590.636247] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 590.638986] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1bf9ccae-8af3-42d0-9727-ab632a268df2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.647924] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd78ec6-9ee2-49cc-a026-ff9d66dfdbb4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.671927] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a could not be found. [ 590.672172] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 590.672416] env[61243]: INFO nova.compute.manager [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 590.672593] env[61243]: DEBUG oslo.service.loopingcall [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 590.674976] env[61243]: DEBUG nova.compute.manager [-] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 590.675085] env[61243]: DEBUG nova.network.neutron [-] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 590.690604] env[61243]: DEBUG nova.network.neutron [-] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.988857] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-026beaa8-2fff-4e21-a50a-e97cdb808dff {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.997609] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8590b8-9f2c-4f4a-bb0c-7fe167e01871 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.029503] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e1e595-92e4-4d93-b50e-041ffb102929 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.037163] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d3c0f11-d1ce-455e-bacb-7348063b5491 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.051125] env[61243]: DEBUG nova.compute.provider_tree [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.194388] env[61243]: DEBUG nova.network.neutron [-] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.553420] env[61243]: DEBUG nova.scheduler.client.report [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 591.656025] env[61243]: INFO nova.scheduler.client.report [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Deleted allocations for instance f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d [ 591.696942] env[61243]: INFO nova.compute.manager [-] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Took 1.02 seconds to deallocate network for instance. [ 591.700852] env[61243]: DEBUG nova.compute.claims [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 591.701462] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.057929] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.540s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.058442] env[61243]: DEBUG nova.compute.manager [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 592.061256] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 42.702s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.164350] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d62aa01-8c13-40ca-95ea-5bad6fa8def9 tempest-ServersTestJSON-878488634 tempest-ServersTestJSON-878488634-project-member] Lock "f7f5e05e-4191-4f3c-90a5-aa4f3f02fc9d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.657s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.566705] env[61243]: DEBUG nova.compute.utils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 592.571264] env[61243]: DEBUG nova.compute.manager [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 592.571467] env[61243]: DEBUG nova.network.neutron [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 592.637504] env[61243]: DEBUG nova.policy [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd87217ad30904ba99afe50c7a5f895b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d9ee1fd63864856a65468a4059161a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 592.669290] env[61243]: DEBUG nova.compute.manager [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 593.078179] env[61243]: DEBUG nova.compute.manager [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 593.094217] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dadaa055-ea08-48f9-9210-69fc90eb73ab {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.101072] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c548ba3-e992-48d6-8f19-9020b791e8ce {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.131727] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db483dad-56b8-4ea5-9cb9-1768378d12f0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.136043] env[61243]: DEBUG nova.network.neutron [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Successfully created port: d8ba96aa-f0b8-463f-bf1e-2e868a893c04 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 593.141493] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa107472-edb3-4f18-b5aa-596eabe8cbee {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.156434] env[61243]: DEBUG nova.compute.provider_tree [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.193591] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.660263] env[61243]: DEBUG nova.scheduler.client.report [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.091989] env[61243]: DEBUG nova.compute.manager [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 594.102474] env[61243]: DEBUG nova.compute.manager [req-df54de31-382a-42b6-a4de-6e1d9bf2b780 req-99dbc282-412b-4d2d-87d1-124a21fa0194 service nova] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Received event network-changed-d8ba96aa-f0b8-463f-bf1e-2e868a893c04 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 594.103677] env[61243]: DEBUG nova.compute.manager [req-df54de31-382a-42b6-a4de-6e1d9bf2b780 req-99dbc282-412b-4d2d-87d1-124a21fa0194 service nova] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Refreshing instance network info cache due to event network-changed-d8ba96aa-f0b8-463f-bf1e-2e868a893c04. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 594.103677] env[61243]: DEBUG oslo_concurrency.lockutils [req-df54de31-382a-42b6-a4de-6e1d9bf2b780 req-99dbc282-412b-4d2d-87d1-124a21fa0194 service nova] Acquiring lock "refresh_cache-762cbca7-4729-458a-b2af-2978723a25dc" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.103677] env[61243]: DEBUG oslo_concurrency.lockutils [req-df54de31-382a-42b6-a4de-6e1d9bf2b780 req-99dbc282-412b-4d2d-87d1-124a21fa0194 service nova] Acquired lock "refresh_cache-762cbca7-4729-458a-b2af-2978723a25dc" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.103677] env[61243]: DEBUG nova.network.neutron [req-df54de31-382a-42b6-a4de-6e1d9bf2b780 req-99dbc282-412b-4d2d-87d1-124a21fa0194 service nova] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Refreshing network info cache for port d8ba96aa-f0b8-463f-bf1e-2e868a893c04 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 594.128320] env[61243]: DEBUG nova.virt.hardware [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:03:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='13201264',id=23,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1302718257',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 594.128320] env[61243]: DEBUG nova.virt.hardware [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 594.128538] env[61243]: DEBUG nova.virt.hardware [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 594.128616] env[61243]: DEBUG nova.virt.hardware [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 594.128761] env[61243]: DEBUG nova.virt.hardware [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 594.128906] env[61243]: DEBUG nova.virt.hardware [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 594.129123] env[61243]: DEBUG nova.virt.hardware [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 594.129362] env[61243]: DEBUG nova.virt.hardware [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 594.129545] env[61243]: DEBUG nova.virt.hardware [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 594.129742] env[61243]: DEBUG nova.virt.hardware [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 594.129921] env[61243]: DEBUG nova.virt.hardware [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 594.130797] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed9a30c-e447-4699-ab88-22be5f20bafc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.139660] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e272188-4786-48c9-a97d-33fc20aea1a2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.165427] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.104s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.166045] env[61243]: ERROR nova.compute.manager [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b2bb5756-8638-4396-acc4-0b333a9578c9, please check neutron logs for more information. [ 594.166045] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Traceback (most recent call last): [ 594.166045] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 594.166045] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] self.driver.spawn(context, instance, image_meta, [ 594.166045] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 594.166045] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.166045] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.166045] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] vm_ref = self.build_virtual_machine(instance, [ 594.166045] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.166045] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.166045] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.166526] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] for vif in network_info: [ 594.166526] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.166526] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] return self._sync_wrapper(fn, *args, **kwargs) [ 594.166526] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.166526] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] self.wait() [ 594.166526] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.166526] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] self[:] = self._gt.wait() [ 594.166526] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.166526] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] return self._exit_event.wait() [ 594.166526] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.166526] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] result = hub.switch() [ 594.166526] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.166526] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] return self.greenlet.switch() [ 594.167858] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.167858] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] result = function(*args, **kwargs) [ 594.167858] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 594.167858] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] return func(*args, **kwargs) [ 594.167858] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.167858] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] raise e [ 594.167858] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.167858] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] nwinfo = self.network_api.allocate_for_instance( [ 594.167858] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.167858] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] created_port_ids = self._update_ports_for_instance( [ 594.167858] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.167858] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] with excutils.save_and_reraise_exception(): [ 594.167858] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.168268] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] self.force_reraise() [ 594.168268] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.168268] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] raise self.value [ 594.168268] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.168268] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] updated_port = self._update_port( [ 594.168268] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.168268] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] _ensure_no_port_binding_failure(port) [ 594.168268] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.168268] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] raise exception.PortBindingFailed(port_id=port['id']) [ 594.168268] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] nova.exception.PortBindingFailed: Binding failed for port b2bb5756-8638-4396-acc4-0b333a9578c9, please check neutron logs for more information. [ 594.168268] env[61243]: ERROR nova.compute.manager [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] [ 594.168633] env[61243]: DEBUG nova.compute.utils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Binding failed for port b2bb5756-8638-4396-acc4-0b333a9578c9, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 594.168633] env[61243]: DEBUG nova.compute.manager [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Build of instance 25a896cb-fd91-4f16-a8cd-6de58fcfcc27 was re-scheduled: Binding failed for port b2bb5756-8638-4396-acc4-0b333a9578c9, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 594.169063] env[61243]: DEBUG nova.compute.manager [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 594.169294] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Acquiring lock "refresh_cache-25a896cb-fd91-4f16-a8cd-6de58fcfcc27" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.169440] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Acquired lock "refresh_cache-25a896cb-fd91-4f16-a8cd-6de58fcfcc27" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.169596] env[61243]: DEBUG nova.network.neutron [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 594.171246] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 40.134s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.249600] env[61243]: ERROR nova.compute.manager [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d8ba96aa-f0b8-463f-bf1e-2e868a893c04, please check neutron logs for more information. [ 594.249600] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 594.249600] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.249600] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 594.249600] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.249600] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 594.249600] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.249600] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 594.249600] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.249600] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 594.249600] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.249600] env[61243]: ERROR nova.compute.manager raise self.value [ 594.249600] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.249600] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 594.249600] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.249600] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 594.250173] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.250173] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 594.250173] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d8ba96aa-f0b8-463f-bf1e-2e868a893c04, please check neutron logs for more information. [ 594.250173] env[61243]: ERROR nova.compute.manager [ 594.250173] env[61243]: Traceback (most recent call last): [ 594.250173] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 594.250173] env[61243]: listener.cb(fileno) [ 594.250173] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.250173] env[61243]: result = function(*args, **kwargs) [ 594.250173] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 594.250173] env[61243]: return func(*args, **kwargs) [ 594.250173] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.250173] env[61243]: raise e [ 594.250173] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.250173] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 594.250173] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.250173] env[61243]: created_port_ids = self._update_ports_for_instance( [ 594.250173] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.250173] env[61243]: with excutils.save_and_reraise_exception(): [ 594.250173] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.250173] env[61243]: self.force_reraise() [ 594.250173] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.250173] env[61243]: raise self.value [ 594.250173] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.250173] env[61243]: updated_port = self._update_port( [ 594.250173] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.250173] env[61243]: _ensure_no_port_binding_failure(port) [ 594.250173] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.250173] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 594.250985] env[61243]: nova.exception.PortBindingFailed: Binding failed for port d8ba96aa-f0b8-463f-bf1e-2e868a893c04, please check neutron logs for more information. [ 594.250985] env[61243]: Removing descriptor: 15 [ 594.250985] env[61243]: ERROR nova.compute.manager [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d8ba96aa-f0b8-463f-bf1e-2e868a893c04, please check neutron logs for more information. [ 594.250985] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Traceback (most recent call last): [ 594.250985] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 594.250985] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] yield resources [ 594.250985] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 594.250985] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] self.driver.spawn(context, instance, image_meta, [ 594.250985] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 594.250985] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.250985] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.250985] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] vm_ref = self.build_virtual_machine(instance, [ 594.251372] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.251372] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.251372] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.251372] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] for vif in network_info: [ 594.251372] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.251372] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] return self._sync_wrapper(fn, *args, **kwargs) [ 594.251372] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.251372] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] self.wait() [ 594.251372] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.251372] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] self[:] = self._gt.wait() [ 594.251372] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.251372] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] return self._exit_event.wait() [ 594.251372] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.251768] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] result = hub.switch() [ 594.251768] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.251768] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] return self.greenlet.switch() [ 594.251768] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.251768] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] result = function(*args, **kwargs) [ 594.251768] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 594.251768] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] return func(*args, **kwargs) [ 594.251768] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.251768] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] raise e [ 594.251768] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.251768] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] nwinfo = self.network_api.allocate_for_instance( [ 594.251768] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.251768] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] created_port_ids = self._update_ports_for_instance( [ 594.252163] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.252163] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] with excutils.save_and_reraise_exception(): [ 594.252163] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.252163] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] self.force_reraise() [ 594.252163] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.252163] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] raise self.value [ 594.252163] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.252163] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] updated_port = self._update_port( [ 594.252163] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.252163] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] _ensure_no_port_binding_failure(port) [ 594.252163] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.252163] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] raise exception.PortBindingFailed(port_id=port['id']) [ 594.252532] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] nova.exception.PortBindingFailed: Binding failed for port d8ba96aa-f0b8-463f-bf1e-2e868a893c04, please check neutron logs for more information. [ 594.252532] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] [ 594.252532] env[61243]: INFO nova.compute.manager [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Terminating instance [ 594.253964] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Acquiring lock "refresh_cache-762cbca7-4729-458a-b2af-2978723a25dc" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.625112] env[61243]: DEBUG nova.network.neutron [req-df54de31-382a-42b6-a4de-6e1d9bf2b780 req-99dbc282-412b-4d2d-87d1-124a21fa0194 service nova] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.705641] env[61243]: DEBUG nova.network.neutron [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.855883] env[61243]: DEBUG nova.network.neutron [req-df54de31-382a-42b6-a4de-6e1d9bf2b780 req-99dbc282-412b-4d2d-87d1-124a21fa0194 service nova] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.902156] env[61243]: DEBUG nova.network.neutron [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.116706] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ffae13-a89d-4af7-b134-fe6af12cf593 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.124306] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d6c46f6-2186-4479-b17b-59269a86a1db {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.155009] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8154669e-8f8e-4376-8a80-749a671ce827 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.163373] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e838ea3b-45d0-4947-bda8-efb89f344dd3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.177540] env[61243]: DEBUG nova.compute.provider_tree [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.359462] env[61243]: DEBUG oslo_concurrency.lockutils [req-df54de31-382a-42b6-a4de-6e1d9bf2b780 req-99dbc282-412b-4d2d-87d1-124a21fa0194 service nova] Releasing lock "refresh_cache-762cbca7-4729-458a-b2af-2978723a25dc" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.359877] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Acquired lock "refresh_cache-762cbca7-4729-458a-b2af-2978723a25dc" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.360207] env[61243]: DEBUG nova.network.neutron [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 595.404906] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Releasing lock "refresh_cache-25a896cb-fd91-4f16-a8cd-6de58fcfcc27" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.404906] env[61243]: DEBUG nova.compute.manager [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 595.404906] env[61243]: DEBUG nova.compute.manager [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 595.404906] env[61243]: DEBUG nova.network.neutron [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 595.420930] env[61243]: DEBUG nova.network.neutron [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.681349] env[61243]: DEBUG nova.scheduler.client.report [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.877897] env[61243]: DEBUG nova.network.neutron [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.923451] env[61243]: DEBUG nova.network.neutron [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.949267] env[61243]: DEBUG nova.network.neutron [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.150796] env[61243]: DEBUG nova.compute.manager [req-7e3892f0-84bb-4b2d-99e3-ac38062f984d req-86ee78d7-5b10-4bf8-b214-aa92eb14a6c8 service nova] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Received event network-vif-deleted-d8ba96aa-f0b8-463f-bf1e-2e868a893c04 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 596.186513] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.015s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.187249] env[61243]: ERROR nova.compute.manager [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4a04ed28-25ad-4709-a762-c930986c585e, please check neutron logs for more information. [ 596.187249] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Traceback (most recent call last): [ 596.187249] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 596.187249] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] self.driver.spawn(context, instance, image_meta, [ 596.187249] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 596.187249] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.187249] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.187249] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] vm_ref = self.build_virtual_machine(instance, [ 596.187249] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.187249] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.187249] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.187723] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] for vif in network_info: [ 596.187723] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 596.187723] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] return self._sync_wrapper(fn, *args, **kwargs) [ 596.187723] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 596.187723] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] self.wait() [ 596.187723] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 596.187723] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] self[:] = self._gt.wait() [ 596.187723] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.187723] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] return self._exit_event.wait() [ 596.187723] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.187723] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] result = hub.switch() [ 596.187723] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.187723] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] return self.greenlet.switch() [ 596.188331] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.188331] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] result = function(*args, **kwargs) [ 596.188331] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 596.188331] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] return func(*args, **kwargs) [ 596.188331] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.188331] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] raise e [ 596.188331] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.188331] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] nwinfo = self.network_api.allocate_for_instance( [ 596.188331] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.188331] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] created_port_ids = self._update_ports_for_instance( [ 596.188331] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.188331] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] with excutils.save_and_reraise_exception(): [ 596.188331] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.188759] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] self.force_reraise() [ 596.188759] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.188759] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] raise self.value [ 596.188759] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.188759] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] updated_port = self._update_port( [ 596.188759] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.188759] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] _ensure_no_port_binding_failure(port) [ 596.188759] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.188759] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] raise exception.PortBindingFailed(port_id=port['id']) [ 596.188759] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] nova.exception.PortBindingFailed: Binding failed for port 4a04ed28-25ad-4709-a762-c930986c585e, please check neutron logs for more information. [ 596.188759] env[61243]: ERROR nova.compute.manager [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] [ 596.189150] env[61243]: DEBUG nova.compute.utils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Binding failed for port 4a04ed28-25ad-4709-a762-c930986c585e, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 596.189572] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 41.926s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.190750] env[61243]: INFO nova.compute.claims [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 596.193505] env[61243]: DEBUG nova.compute.manager [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Build of instance 8c93ef89-5d27-48d3-8c44-79eb23d66ee3 was re-scheduled: Binding failed for port 4a04ed28-25ad-4709-a762-c930986c585e, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 596.193948] env[61243]: DEBUG nova.compute.manager [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 596.194187] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Acquiring lock "refresh_cache-8c93ef89-5d27-48d3-8c44-79eb23d66ee3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.194338] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Acquired lock "refresh_cache-8c93ef89-5d27-48d3-8c44-79eb23d66ee3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.194496] env[61243]: DEBUG nova.network.neutron [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 596.426980] env[61243]: INFO nova.compute.manager [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] [instance: 25a896cb-fd91-4f16-a8cd-6de58fcfcc27] Took 1.02 seconds to deallocate network for instance. [ 596.453668] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Releasing lock "refresh_cache-762cbca7-4729-458a-b2af-2978723a25dc" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.454172] env[61243]: DEBUG nova.compute.manager [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 596.454380] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 596.454688] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8482c99c-0d73-4975-85f9-1e4c4e8cacf1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.464105] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-229c5cac-d543-4cf8-8b1c-759b975d7031 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.485261] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 762cbca7-4729-458a-b2af-2978723a25dc could not be found. [ 596.485485] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 596.485668] env[61243]: INFO nova.compute.manager [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Took 0.03 seconds to destroy the instance on the hypervisor. [ 596.486046] env[61243]: DEBUG oslo.service.loopingcall [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 596.486770] env[61243]: DEBUG nova.compute.manager [-] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 596.486867] env[61243]: DEBUG nova.network.neutron [-] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 596.503772] env[61243]: DEBUG nova.network.neutron [-] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.714726] env[61243]: DEBUG nova.network.neutron [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.790902] env[61243]: DEBUG nova.network.neutron [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.007371] env[61243]: DEBUG nova.network.neutron [-] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.296876] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Releasing lock "refresh_cache-8c93ef89-5d27-48d3-8c44-79eb23d66ee3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.297997] env[61243]: DEBUG nova.compute.manager [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 597.297997] env[61243]: DEBUG nova.compute.manager [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 597.297997] env[61243]: DEBUG nova.network.neutron [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 597.313683] env[61243]: DEBUG nova.network.neutron [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.461919] env[61243]: INFO nova.scheduler.client.report [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Deleted allocations for instance 25a896cb-fd91-4f16-a8cd-6de58fcfcc27 [ 597.509467] env[61243]: INFO nova.compute.manager [-] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Took 1.02 seconds to deallocate network for instance. [ 597.515667] env[61243]: DEBUG nova.compute.claims [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 597.515868] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.651407] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c76c682f-a0eb-45ca-b549-4bb471ff6822 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.658660] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b229868-8324-4762-ac7b-4a01c8317b55 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.687183] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ead158-5757-40d5-b999-e7676dd8f5cb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.694505] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53c4c012-f524-4da1-b07a-8cfbc455f4f4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.707747] env[61243]: DEBUG nova.compute.provider_tree [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.819022] env[61243]: DEBUG nova.network.neutron [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.971533] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9f4da82-276b-40d3-a109-244a2a12420d tempest-ServerExternalEventsTest-1022624430 tempest-ServerExternalEventsTest-1022624430-project-member] Lock "25a896cb-fd91-4f16-a8cd-6de58fcfcc27" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.448s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.211850] env[61243]: DEBUG nova.scheduler.client.report [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 598.324473] env[61243]: INFO nova.compute.manager [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 8c93ef89-5d27-48d3-8c44-79eb23d66ee3] Took 1.03 seconds to deallocate network for instance. [ 598.474769] env[61243]: DEBUG nova.compute.manager [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 598.719908] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.531s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.720430] env[61243]: DEBUG nova.compute.manager [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 598.723073] env[61243]: DEBUG oslo_concurrency.lockutils [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 38.902s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.723268] env[61243]: DEBUG nova.objects.instance [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61243) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 599.005171] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.227821] env[61243]: DEBUG nova.compute.utils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 599.232999] env[61243]: DEBUG nova.compute.manager [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 599.237392] env[61243]: DEBUG nova.network.neutron [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 599.289951] env[61243]: DEBUG nova.policy [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1d99c406cdc34aca9344b9ffea64b1fe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aae25749b1d44c59afff61a8ea74c803', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 599.355391] env[61243]: INFO nova.scheduler.client.report [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Deleted allocations for instance 8c93ef89-5d27-48d3-8c44-79eb23d66ee3 [ 599.592292] env[61243]: DEBUG nova.network.neutron [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Successfully created port: 92c79f78-3449-4425-a9d2-a2df126769b9 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 599.733895] env[61243]: DEBUG nova.compute.manager [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 599.746076] env[61243]: DEBUG oslo_concurrency.lockutils [None req-79ea0fd8-2318-4c73-9a4a-b568ab04d6c0 tempest-ServersAdmin275Test-453172524 tempest-ServersAdmin275Test-453172524-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.023s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.747962] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 39.830s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.749325] env[61243]: INFO nova.compute.claims [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.863909] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9694af40-b5ae-472e-b46d-11a5af845798 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Lock "8c93ef89-5d27-48d3-8c44-79eb23d66ee3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.653s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.369681] env[61243]: DEBUG nova.compute.manager [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 600.729964] env[61243]: DEBUG nova.compute.manager [req-3a066b17-85a4-4d1c-a0de-bd14cca6db5d req-1fb88907-33fc-4e94-8308-aab882ef9591 service nova] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Received event network-changed-92c79f78-3449-4425-a9d2-a2df126769b9 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.730210] env[61243]: DEBUG nova.compute.manager [req-3a066b17-85a4-4d1c-a0de-bd14cca6db5d req-1fb88907-33fc-4e94-8308-aab882ef9591 service nova] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Refreshing instance network info cache due to event network-changed-92c79f78-3449-4425-a9d2-a2df126769b9. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 600.730883] env[61243]: DEBUG oslo_concurrency.lockutils [req-3a066b17-85a4-4d1c-a0de-bd14cca6db5d req-1fb88907-33fc-4e94-8308-aab882ef9591 service nova] Acquiring lock "refresh_cache-3700deef-a438-4920-b35f-8fba370a5984" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.731010] env[61243]: DEBUG oslo_concurrency.lockutils [req-3a066b17-85a4-4d1c-a0de-bd14cca6db5d req-1fb88907-33fc-4e94-8308-aab882ef9591 service nova] Acquired lock "refresh_cache-3700deef-a438-4920-b35f-8fba370a5984" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.731195] env[61243]: DEBUG nova.network.neutron [req-3a066b17-85a4-4d1c-a0de-bd14cca6db5d req-1fb88907-33fc-4e94-8308-aab882ef9591 service nova] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Refreshing network info cache for port 92c79f78-3449-4425-a9d2-a2df126769b9 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 600.745518] env[61243]: DEBUG nova.compute.manager [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 600.774241] env[61243]: DEBUG nova.virt.hardware [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 600.774382] env[61243]: DEBUG nova.virt.hardware [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 600.775061] env[61243]: DEBUG nova.virt.hardware [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 600.775061] env[61243]: DEBUG nova.virt.hardware [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 600.775061] env[61243]: DEBUG nova.virt.hardware [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 600.775414] env[61243]: DEBUG nova.virt.hardware [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 600.775839] env[61243]: DEBUG nova.virt.hardware [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 600.776042] env[61243]: DEBUG nova.virt.hardware [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 600.776356] env[61243]: DEBUG nova.virt.hardware [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 600.776605] env[61243]: DEBUG nova.virt.hardware [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 600.776864] env[61243]: DEBUG nova.virt.hardware [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 600.778105] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e9a8c2d-e7bc-4b3d-9138-cf85eed2c842 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.788605] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acdabd37-3eb1-4eaf-9b87-129b11e1e130 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.890229] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.903679] env[61243]: ERROR nova.compute.manager [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 92c79f78-3449-4425-a9d2-a2df126769b9, please check neutron logs for more information. [ 600.903679] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 600.903679] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.903679] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 600.903679] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.903679] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 600.903679] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.903679] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 600.903679] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.903679] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 600.903679] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.903679] env[61243]: ERROR nova.compute.manager raise self.value [ 600.903679] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.903679] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 600.903679] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.903679] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 600.904121] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.904121] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 600.904121] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 92c79f78-3449-4425-a9d2-a2df126769b9, please check neutron logs for more information. [ 600.904121] env[61243]: ERROR nova.compute.manager [ 600.904121] env[61243]: Traceback (most recent call last): [ 600.904121] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 600.904121] env[61243]: listener.cb(fileno) [ 600.904121] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.904121] env[61243]: result = function(*args, **kwargs) [ 600.904121] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 600.904121] env[61243]: return func(*args, **kwargs) [ 600.904121] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.904121] env[61243]: raise e [ 600.904121] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.904121] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 600.904121] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.904121] env[61243]: created_port_ids = self._update_ports_for_instance( [ 600.904121] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.904121] env[61243]: with excutils.save_and_reraise_exception(): [ 600.904121] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.904121] env[61243]: self.force_reraise() [ 600.904121] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.904121] env[61243]: raise self.value [ 600.904121] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.904121] env[61243]: updated_port = self._update_port( [ 600.904121] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.904121] env[61243]: _ensure_no_port_binding_failure(port) [ 600.904121] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.904121] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 600.904796] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 92c79f78-3449-4425-a9d2-a2df126769b9, please check neutron logs for more information. [ 600.904796] env[61243]: Removing descriptor: 16 [ 600.904796] env[61243]: ERROR nova.compute.manager [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 92c79f78-3449-4425-a9d2-a2df126769b9, please check neutron logs for more information. [ 600.904796] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] Traceback (most recent call last): [ 600.904796] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 600.904796] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] yield resources [ 600.904796] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 600.904796] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] self.driver.spawn(context, instance, image_meta, [ 600.904796] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 600.904796] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.904796] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.904796] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] vm_ref = self.build_virtual_machine(instance, [ 600.905044] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.905044] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.905044] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.905044] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] for vif in network_info: [ 600.905044] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 600.905044] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] return self._sync_wrapper(fn, *args, **kwargs) [ 600.905044] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 600.905044] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] self.wait() [ 600.905044] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 600.905044] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] self[:] = self._gt.wait() [ 600.905044] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.905044] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] return self._exit_event.wait() [ 600.905044] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.905406] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] result = hub.switch() [ 600.905406] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.905406] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] return self.greenlet.switch() [ 600.905406] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.905406] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] result = function(*args, **kwargs) [ 600.905406] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 600.905406] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] return func(*args, **kwargs) [ 600.905406] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.905406] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] raise e [ 600.905406] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.905406] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] nwinfo = self.network_api.allocate_for_instance( [ 600.905406] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.905406] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] created_port_ids = self._update_ports_for_instance( [ 600.905704] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.905704] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] with excutils.save_and_reraise_exception(): [ 600.905704] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.905704] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] self.force_reraise() [ 600.905704] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.905704] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] raise self.value [ 600.905704] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.905704] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] updated_port = self._update_port( [ 600.905704] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.905704] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] _ensure_no_port_binding_failure(port) [ 600.905704] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.905704] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] raise exception.PortBindingFailed(port_id=port['id']) [ 600.905936] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] nova.exception.PortBindingFailed: Binding failed for port 92c79f78-3449-4425-a9d2-a2df126769b9, please check neutron logs for more information. [ 600.905936] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] [ 600.905936] env[61243]: INFO nova.compute.manager [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Terminating instance [ 600.907784] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Acquiring lock "refresh_cache-3700deef-a438-4920-b35f-8fba370a5984" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.251796] env[61243]: DEBUG nova.network.neutron [req-3a066b17-85a4-4d1c-a0de-bd14cca6db5d req-1fb88907-33fc-4e94-8308-aab882ef9591 service nova] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.255783] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5f8ef7-4494-4624-83a5-132c92c034da {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.266021] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb29f255-5305-4614-91f5-c553714ecd97 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.303399] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab275ec5-458f-4e90-bc30-fa776a1867e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.312855] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb473359-aa42-446e-8b40-35214487f0a6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.328803] env[61243]: DEBUG nova.compute.provider_tree [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.344102] env[61243]: DEBUG nova.network.neutron [req-3a066b17-85a4-4d1c-a0de-bd14cca6db5d req-1fb88907-33fc-4e94-8308-aab882ef9591 service nova] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.834630] env[61243]: DEBUG nova.scheduler.client.report [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.844777] env[61243]: DEBUG oslo_concurrency.lockutils [req-3a066b17-85a4-4d1c-a0de-bd14cca6db5d req-1fb88907-33fc-4e94-8308-aab882ef9591 service nova] Releasing lock "refresh_cache-3700deef-a438-4920-b35f-8fba370a5984" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.845583] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Acquired lock "refresh_cache-3700deef-a438-4920-b35f-8fba370a5984" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.846169] env[61243]: DEBUG nova.network.neutron [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.339967] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.592s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.340747] env[61243]: DEBUG nova.compute.manager [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 602.343615] env[61243]: DEBUG oslo_concurrency.lockutils [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 37.513s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.343882] env[61243]: DEBUG oslo_concurrency.lockutils [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.345820] env[61243]: DEBUG oslo_concurrency.lockutils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 36.627s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.347200] env[61243]: INFO nova.compute.claims [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 602.370422] env[61243]: DEBUG nova.network.neutron [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.388259] env[61243]: INFO nova.scheduler.client.report [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Deleted allocations for instance 1b1b2bac-67af-4921-a49a-429017ef7a3b [ 602.463914] env[61243]: DEBUG nova.network.neutron [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.856019] env[61243]: DEBUG nova.compute.utils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 602.856019] env[61243]: DEBUG nova.compute.manager [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 602.856019] env[61243]: DEBUG nova.network.neutron [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 602.908725] env[61243]: DEBUG oslo_concurrency.lockutils [None req-32631473-9f5a-4135-a0d2-1efb52299e53 tempest-ServersAdmin275Test-1249737397 tempest-ServersAdmin275Test-1249737397-project-member] Lock "1b1b2bac-67af-4921-a49a-429017ef7a3b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.889s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.916967] env[61243]: DEBUG nova.policy [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b527d68531a44d5f86baf948988dbdd5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e9d66eafbd54d7b94212eb643876c9d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 602.969923] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Releasing lock "refresh_cache-3700deef-a438-4920-b35f-8fba370a5984" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.969923] env[61243]: DEBUG nova.compute.manager [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 602.969923] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 602.969923] env[61243]: DEBUG nova.compute.manager [req-7548eece-1504-4b31-bd5d-8420f3fb3bbd req-7099960a-7e6a-4894-8ab0-05a4d6a7f6a1 service nova] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Received event network-vif-deleted-92c79f78-3449-4425-a9d2-a2df126769b9 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 602.970600] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7a9d2ec4-6b80-4258-86af-47e9e53bd7db {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.982711] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dbf7e1c-d6cf-4baf-bdf4-8e1c9ea0eeb0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.012759] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3700deef-a438-4920-b35f-8fba370a5984 could not be found. [ 603.013032] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 603.013241] env[61243]: INFO nova.compute.manager [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Took 0.04 seconds to destroy the instance on the hypervisor. [ 603.014557] env[61243]: DEBUG oslo.service.loopingcall [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 603.014557] env[61243]: DEBUG nova.compute.manager [-] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.014557] env[61243]: DEBUG nova.network.neutron [-] [instance: 3700deef-a438-4920-b35f-8fba370a5984] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.036767] env[61243]: DEBUG nova.network.neutron [-] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.198821] env[61243]: DEBUG nova.network.neutron [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Successfully created port: 74fc99ca-72cf-4729-bf32-21f722cbdaf0 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 603.365122] env[61243]: DEBUG nova.compute.manager [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 603.538598] env[61243]: DEBUG nova.network.neutron [-] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.853441] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14defa79-418f-4c7b-85d0-5e502ff4bb0f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.863142] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbdbb26b-97a4-446e-b96d-3cb69d7c6ffa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.901039] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d763ce3b-2fa7-410a-8911-566a74139690 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.910283] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4dafc3e-a027-45ee-97ec-c51bf2888abd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.924353] env[61243]: DEBUG nova.compute.provider_tree [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.047733] env[61243]: INFO nova.compute.manager [-] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Took 1.03 seconds to deallocate network for instance. [ 604.050256] env[61243]: DEBUG nova.compute.claims [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 604.051034] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.292561] env[61243]: ERROR nova.compute.manager [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 74fc99ca-72cf-4729-bf32-21f722cbdaf0, please check neutron logs for more information. [ 604.292561] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 604.292561] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.292561] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 604.292561] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.292561] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 604.292561] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.292561] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 604.292561] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.292561] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 604.292561] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.292561] env[61243]: ERROR nova.compute.manager raise self.value [ 604.292561] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.292561] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 604.292561] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.292561] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 604.292953] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.292953] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 604.292953] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 74fc99ca-72cf-4729-bf32-21f722cbdaf0, please check neutron logs for more information. [ 604.292953] env[61243]: ERROR nova.compute.manager [ 604.292953] env[61243]: Traceback (most recent call last): [ 604.292953] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 604.292953] env[61243]: listener.cb(fileno) [ 604.292953] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.292953] env[61243]: result = function(*args, **kwargs) [ 604.292953] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 604.292953] env[61243]: return func(*args, **kwargs) [ 604.292953] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.292953] env[61243]: raise e [ 604.292953] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.292953] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 604.292953] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.292953] env[61243]: created_port_ids = self._update_ports_for_instance( [ 604.292953] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.292953] env[61243]: with excutils.save_and_reraise_exception(): [ 604.292953] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.292953] env[61243]: self.force_reraise() [ 604.292953] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.292953] env[61243]: raise self.value [ 604.292953] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.292953] env[61243]: updated_port = self._update_port( [ 604.292953] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.292953] env[61243]: _ensure_no_port_binding_failure(port) [ 604.292953] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.292953] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 604.293616] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 74fc99ca-72cf-4729-bf32-21f722cbdaf0, please check neutron logs for more information. [ 604.293616] env[61243]: Removing descriptor: 16 [ 604.409018] env[61243]: DEBUG nova.compute.manager [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 604.430912] env[61243]: DEBUG nova.scheduler.client.report [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.435961] env[61243]: DEBUG nova.virt.hardware [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.436254] env[61243]: DEBUG nova.virt.hardware [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.436417] env[61243]: DEBUG nova.virt.hardware [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.436599] env[61243]: DEBUG nova.virt.hardware [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.437393] env[61243]: DEBUG nova.virt.hardware [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.437393] env[61243]: DEBUG nova.virt.hardware [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.437393] env[61243]: DEBUG nova.virt.hardware [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.437393] env[61243]: DEBUG nova.virt.hardware [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.437573] env[61243]: DEBUG nova.virt.hardware [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.438976] env[61243]: DEBUG nova.virt.hardware [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.438976] env[61243]: DEBUG nova.virt.hardware [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.438976] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f266a97-ec38-49b2-b6af-3d4d74a80443 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.447143] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1378f93c-8b5a-495f-adb2-adfd48d9b188 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.464336] env[61243]: ERROR nova.compute.manager [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 74fc99ca-72cf-4729-bf32-21f722cbdaf0, please check neutron logs for more information. [ 604.464336] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] Traceback (most recent call last): [ 604.464336] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 604.464336] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] yield resources [ 604.464336] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.464336] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] self.driver.spawn(context, instance, image_meta, [ 604.464336] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 604.464336] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.464336] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.464336] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] vm_ref = self.build_virtual_machine(instance, [ 604.464336] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.464636] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.464636] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.464636] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] for vif in network_info: [ 604.464636] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.464636] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] return self._sync_wrapper(fn, *args, **kwargs) [ 604.464636] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.464636] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] self.wait() [ 604.464636] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.464636] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] self[:] = self._gt.wait() [ 604.464636] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.464636] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] return self._exit_event.wait() [ 604.464636] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 604.464636] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] current.throw(*self._exc) [ 604.464904] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.464904] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] result = function(*args, **kwargs) [ 604.464904] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 604.464904] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] return func(*args, **kwargs) [ 604.464904] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.464904] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] raise e [ 604.464904] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.464904] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] nwinfo = self.network_api.allocate_for_instance( [ 604.464904] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.464904] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] created_port_ids = self._update_ports_for_instance( [ 604.464904] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.464904] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] with excutils.save_and_reraise_exception(): [ 604.464904] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.465179] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] self.force_reraise() [ 604.465179] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.465179] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] raise self.value [ 604.465179] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.465179] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] updated_port = self._update_port( [ 604.465179] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.465179] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] _ensure_no_port_binding_failure(port) [ 604.465179] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.465179] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] raise exception.PortBindingFailed(port_id=port['id']) [ 604.465179] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] nova.exception.PortBindingFailed: Binding failed for port 74fc99ca-72cf-4729-bf32-21f722cbdaf0, please check neutron logs for more information. [ 604.465179] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] [ 604.465179] env[61243]: INFO nova.compute.manager [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Terminating instance [ 604.466734] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Acquiring lock "refresh_cache-2451843a-fd32-450e-933e-e635c789834b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.466893] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Acquired lock "refresh_cache-2451843a-fd32-450e-933e-e635c789834b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.467235] env[61243]: DEBUG nova.network.neutron [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 604.943292] env[61243]: DEBUG oslo_concurrency.lockutils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.597s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.944040] env[61243]: DEBUG nova.compute.manager [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 604.951024] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.993s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.951024] env[61243]: INFO nova.compute.claims [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 605.000573] env[61243]: DEBUG nova.network.neutron [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.136751] env[61243]: DEBUG nova.network.neutron [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.223667] env[61243]: DEBUG nova.compute.manager [req-3f13cc8a-2997-47ab-b373-d6bf747b9ad1 req-0735bf2f-c433-4a1b-bc7b-dbf441ef152d service nova] [instance: 2451843a-fd32-450e-933e-e635c789834b] Received event network-changed-74fc99ca-72cf-4729-bf32-21f722cbdaf0 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 605.225171] env[61243]: DEBUG nova.compute.manager [req-3f13cc8a-2997-47ab-b373-d6bf747b9ad1 req-0735bf2f-c433-4a1b-bc7b-dbf441ef152d service nova] [instance: 2451843a-fd32-450e-933e-e635c789834b] Refreshing instance network info cache due to event network-changed-74fc99ca-72cf-4729-bf32-21f722cbdaf0. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 605.225381] env[61243]: DEBUG oslo_concurrency.lockutils [req-3f13cc8a-2997-47ab-b373-d6bf747b9ad1 req-0735bf2f-c433-4a1b-bc7b-dbf441ef152d service nova] Acquiring lock "refresh_cache-2451843a-fd32-450e-933e-e635c789834b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.449091] env[61243]: DEBUG nova.compute.utils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 605.450578] env[61243]: DEBUG nova.compute.manager [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 605.450654] env[61243]: DEBUG nova.network.neutron [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 605.505015] env[61243]: DEBUG nova.policy [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf35957b84954d238bd75756426033e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06c0bbb43a964f6fa145cfaa06ce35ac', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 605.639095] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Releasing lock "refresh_cache-2451843a-fd32-450e-933e-e635c789834b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.639474] env[61243]: DEBUG nova.compute.manager [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 605.639713] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 605.640128] env[61243]: DEBUG oslo_concurrency.lockutils [req-3f13cc8a-2997-47ab-b373-d6bf747b9ad1 req-0735bf2f-c433-4a1b-bc7b-dbf441ef152d service nova] Acquired lock "refresh_cache-2451843a-fd32-450e-933e-e635c789834b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.640409] env[61243]: DEBUG nova.network.neutron [req-3f13cc8a-2997-47ab-b373-d6bf747b9ad1 req-0735bf2f-c433-4a1b-bc7b-dbf441ef152d service nova] [instance: 2451843a-fd32-450e-933e-e635c789834b] Refreshing network info cache for port 74fc99ca-72cf-4729-bf32-21f722cbdaf0 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 605.641552] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f91911e4-c13d-4b3b-baa6-d2cd9a0689c7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.651520] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dceaec0-ce65-4f62-9664-9a7010bb57da {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.675985] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2451843a-fd32-450e-933e-e635c789834b could not be found. [ 605.676231] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 605.676412] env[61243]: INFO nova.compute.manager [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 605.676649] env[61243]: DEBUG oslo.service.loopingcall [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 605.676858] env[61243]: DEBUG nova.compute.manager [-] [instance: 2451843a-fd32-450e-933e-e635c789834b] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.676951] env[61243]: DEBUG nova.network.neutron [-] [instance: 2451843a-fd32-450e-933e-e635c789834b] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 605.702117] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Acquiring lock "61d68f1c-6a48-475a-8cf5-fa6ce3b09f20" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.702364] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Lock "61d68f1c-6a48-475a-8cf5-fa6ce3b09f20" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.838027] env[61243]: DEBUG nova.network.neutron [-] [instance: 2451843a-fd32-450e-933e-e635c789834b] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.953659] env[61243]: DEBUG nova.compute.manager [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 606.130657] env[61243]: DEBUG nova.network.neutron [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Successfully created port: a5db5193-898e-47cb-8133-641566a114af {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 606.166543] env[61243]: DEBUG nova.network.neutron [req-3f13cc8a-2997-47ab-b373-d6bf747b9ad1 req-0735bf2f-c433-4a1b-bc7b-dbf441ef152d service nova] [instance: 2451843a-fd32-450e-933e-e635c789834b] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.340778] env[61243]: DEBUG nova.network.neutron [-] [instance: 2451843a-fd32-450e-933e-e635c789834b] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.353244] env[61243]: DEBUG nova.network.neutron [req-3f13cc8a-2997-47ab-b373-d6bf747b9ad1 req-0735bf2f-c433-4a1b-bc7b-dbf441ef152d service nova] [instance: 2451843a-fd32-450e-933e-e635c789834b] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.421684] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f92ad72-3bf0-424b-b82a-d22154e46d9a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.430927] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516adefb-7234-4061-9022-e6cb6e0bdfab {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.469176] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88b902ac-3707-4a22-bc4f-be8981bf85b0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.477582] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c894e8df-4cb4-4dc5-8e35-303459a42626 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.496096] env[61243]: DEBUG nova.compute.provider_tree [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.843499] env[61243]: INFO nova.compute.manager [-] [instance: 2451843a-fd32-450e-933e-e635c789834b] Took 1.17 seconds to deallocate network for instance. [ 606.848124] env[61243]: DEBUG nova.compute.claims [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 606.848124] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.855667] env[61243]: DEBUG oslo_concurrency.lockutils [req-3f13cc8a-2997-47ab-b373-d6bf747b9ad1 req-0735bf2f-c433-4a1b-bc7b-dbf441ef152d service nova] Releasing lock "refresh_cache-2451843a-fd32-450e-933e-e635c789834b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.855911] env[61243]: DEBUG nova.compute.manager [req-3f13cc8a-2997-47ab-b373-d6bf747b9ad1 req-0735bf2f-c433-4a1b-bc7b-dbf441ef152d service nova] [instance: 2451843a-fd32-450e-933e-e635c789834b] Received event network-vif-deleted-74fc99ca-72cf-4729-bf32-21f722cbdaf0 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 606.974237] env[61243]: DEBUG nova.compute.manager [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 606.999391] env[61243]: DEBUG nova.scheduler.client.report [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 607.015210] env[61243]: DEBUG nova.virt.hardware [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 607.015281] env[61243]: DEBUG nova.virt.hardware [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 607.015830] env[61243]: DEBUG nova.virt.hardware [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 607.015830] env[61243]: DEBUG nova.virt.hardware [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 607.015830] env[61243]: DEBUG nova.virt.hardware [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 607.015965] env[61243]: DEBUG nova.virt.hardware [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 607.016108] env[61243]: DEBUG nova.virt.hardware [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 607.016285] env[61243]: DEBUG nova.virt.hardware [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 607.016512] env[61243]: DEBUG nova.virt.hardware [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 607.018042] env[61243]: DEBUG nova.virt.hardware [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 607.018042] env[61243]: DEBUG nova.virt.hardware [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 607.018042] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c4a1745-fef9-477b-b3b0-a4cb9d07890d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.028854] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b1fdd3-d85b-4fe8-8940-a1e15ee868fa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.508386] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.560s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.509420] env[61243]: DEBUG nova.compute.manager [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 607.514171] env[61243]: DEBUG oslo_concurrency.lockutils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.056s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.518871] env[61243]: INFO nova.compute.claims [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 607.596740] env[61243]: DEBUG nova.compute.manager [req-0f57809c-d428-499d-b993-dc631df09720 req-8125b716-6a53-45a4-8a00-88822e508318 service nova] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Received event network-changed-a5db5193-898e-47cb-8133-641566a114af {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 607.596923] env[61243]: DEBUG nova.compute.manager [req-0f57809c-d428-499d-b993-dc631df09720 req-8125b716-6a53-45a4-8a00-88822e508318 service nova] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Refreshing instance network info cache due to event network-changed-a5db5193-898e-47cb-8133-641566a114af. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 607.597167] env[61243]: DEBUG oslo_concurrency.lockutils [req-0f57809c-d428-499d-b993-dc631df09720 req-8125b716-6a53-45a4-8a00-88822e508318 service nova] Acquiring lock "refresh_cache-f21b7977-747e-437f-b03d-8442ed58756e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.597314] env[61243]: DEBUG oslo_concurrency.lockutils [req-0f57809c-d428-499d-b993-dc631df09720 req-8125b716-6a53-45a4-8a00-88822e508318 service nova] Acquired lock "refresh_cache-f21b7977-747e-437f-b03d-8442ed58756e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.597475] env[61243]: DEBUG nova.network.neutron [req-0f57809c-d428-499d-b993-dc631df09720 req-8125b716-6a53-45a4-8a00-88822e508318 service nova] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Refreshing network info cache for port a5db5193-898e-47cb-8133-641566a114af {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 607.766500] env[61243]: ERROR nova.compute.manager [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a5db5193-898e-47cb-8133-641566a114af, please check neutron logs for more information. [ 607.766500] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 607.766500] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.766500] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 607.766500] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.766500] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 607.766500] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.766500] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 607.766500] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.766500] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 607.766500] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.766500] env[61243]: ERROR nova.compute.manager raise self.value [ 607.766500] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.766500] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 607.766500] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.766500] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 607.766910] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.766910] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 607.766910] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a5db5193-898e-47cb-8133-641566a114af, please check neutron logs for more information. [ 607.766910] env[61243]: ERROR nova.compute.manager [ 607.766910] env[61243]: Traceback (most recent call last): [ 607.766910] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 607.766910] env[61243]: listener.cb(fileno) [ 607.766910] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.766910] env[61243]: result = function(*args, **kwargs) [ 607.766910] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 607.766910] env[61243]: return func(*args, **kwargs) [ 607.766910] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.766910] env[61243]: raise e [ 607.766910] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.766910] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 607.766910] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.766910] env[61243]: created_port_ids = self._update_ports_for_instance( [ 607.766910] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.766910] env[61243]: with excutils.save_and_reraise_exception(): [ 607.766910] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.766910] env[61243]: self.force_reraise() [ 607.766910] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.766910] env[61243]: raise self.value [ 607.766910] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.766910] env[61243]: updated_port = self._update_port( [ 607.766910] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.766910] env[61243]: _ensure_no_port_binding_failure(port) [ 607.766910] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.766910] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 607.767570] env[61243]: nova.exception.PortBindingFailed: Binding failed for port a5db5193-898e-47cb-8133-641566a114af, please check neutron logs for more information. [ 607.767570] env[61243]: Removing descriptor: 16 [ 607.767570] env[61243]: ERROR nova.compute.manager [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a5db5193-898e-47cb-8133-641566a114af, please check neutron logs for more information. [ 607.767570] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] Traceback (most recent call last): [ 607.767570] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 607.767570] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] yield resources [ 607.767570] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 607.767570] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] self.driver.spawn(context, instance, image_meta, [ 607.767570] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 607.767570] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.767570] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.767570] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] vm_ref = self.build_virtual_machine(instance, [ 607.767919] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.767919] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.767919] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.767919] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] for vif in network_info: [ 607.767919] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.767919] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] return self._sync_wrapper(fn, *args, **kwargs) [ 607.767919] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.767919] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] self.wait() [ 607.767919] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.767919] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] self[:] = self._gt.wait() [ 607.767919] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.767919] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] return self._exit_event.wait() [ 607.767919] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 607.768196] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] result = hub.switch() [ 607.768196] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 607.768196] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] return self.greenlet.switch() [ 607.768196] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.768196] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] result = function(*args, **kwargs) [ 607.768196] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 607.768196] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] return func(*args, **kwargs) [ 607.768196] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.768196] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] raise e [ 607.768196] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.768196] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] nwinfo = self.network_api.allocate_for_instance( [ 607.768196] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.768196] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] created_port_ids = self._update_ports_for_instance( [ 607.768514] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.768514] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] with excutils.save_and_reraise_exception(): [ 607.768514] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.768514] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] self.force_reraise() [ 607.768514] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.768514] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] raise self.value [ 607.768514] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.768514] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] updated_port = self._update_port( [ 607.768514] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.768514] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] _ensure_no_port_binding_failure(port) [ 607.768514] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.768514] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] raise exception.PortBindingFailed(port_id=port['id']) [ 607.768820] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] nova.exception.PortBindingFailed: Binding failed for port a5db5193-898e-47cb-8133-641566a114af, please check neutron logs for more information. [ 607.768820] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] [ 607.768820] env[61243]: INFO nova.compute.manager [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Terminating instance [ 607.776341] env[61243]: DEBUG oslo_concurrency.lockutils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquiring lock "refresh_cache-f21b7977-747e-437f-b03d-8442ed58756e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.023620] env[61243]: DEBUG nova.compute.utils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 608.024726] env[61243]: DEBUG nova.compute.manager [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 608.024897] env[61243]: DEBUG nova.network.neutron [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 608.125305] env[61243]: DEBUG nova.network.neutron [req-0f57809c-d428-499d-b993-dc631df09720 req-8125b716-6a53-45a4-8a00-88822e508318 service nova] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.134013] env[61243]: DEBUG nova.policy [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a69c94f04d904f0aa937b26553964558', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55b84457564e4f11bdab80559389b70b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 608.298678] env[61243]: DEBUG nova.network.neutron [req-0f57809c-d428-499d-b993-dc631df09720 req-8125b716-6a53-45a4-8a00-88822e508318 service nova] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.535473] env[61243]: DEBUG nova.compute.manager [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 608.808579] env[61243]: DEBUG oslo_concurrency.lockutils [req-0f57809c-d428-499d-b993-dc631df09720 req-8125b716-6a53-45a4-8a00-88822e508318 service nova] Releasing lock "refresh_cache-f21b7977-747e-437f-b03d-8442ed58756e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.809047] env[61243]: DEBUG oslo_concurrency.lockutils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquired lock "refresh_cache-f21b7977-747e-437f-b03d-8442ed58756e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.809315] env[61243]: DEBUG nova.network.neutron [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 608.854440] env[61243]: DEBUG nova.network.neutron [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Successfully created port: 432beef1-3f34-433c-9998-2c6210f10935 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 609.067041] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e41f24-b0a2-4da8-8c14-de649b101319 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.076493] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1b8efd-c479-4316-9518-0a2ddb9c1c3a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.115514] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8ac417-db6d-4e18-8632-3f2ca7f96f2f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.127041] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Acquiring lock "454d97a2-40d0-4c7a-9df7-04c391715e0d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.127041] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Lock "454d97a2-40d0-4c7a-9df7-04c391715e0d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.127281] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0f40cc6-a839-41d2-ad4a-ba13e46eaa5c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.141829] env[61243]: DEBUG nova.compute.provider_tree [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.330487] env[61243]: DEBUG nova.network.neutron [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.457545] env[61243]: DEBUG nova.network.neutron [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.548475] env[61243]: DEBUG nova.compute.manager [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 609.579380] env[61243]: DEBUG nova.virt.hardware [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 609.579625] env[61243]: DEBUG nova.virt.hardware [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 609.579783] env[61243]: DEBUG nova.virt.hardware [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 609.579996] env[61243]: DEBUG nova.virt.hardware [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 609.580193] env[61243]: DEBUG nova.virt.hardware [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 609.580360] env[61243]: DEBUG nova.virt.hardware [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 609.580569] env[61243]: DEBUG nova.virt.hardware [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 609.580734] env[61243]: DEBUG nova.virt.hardware [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 609.580903] env[61243]: DEBUG nova.virt.hardware [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 609.581078] env[61243]: DEBUG nova.virt.hardware [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 609.581255] env[61243]: DEBUG nova.virt.hardware [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 609.582111] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39ad9213-71d5-4a0b-9aa7-53e75da00675 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.593149] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e878939b-b14f-487f-a6a5-16739d3cdb22 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.642397] env[61243]: DEBUG nova.compute.manager [req-40cc3a8e-d596-447e-a640-ae867c780eb1 req-7aceeaa1-0200-4a48-9989-7728585e409f service nova] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Received event network-vif-deleted-a5db5193-898e-47cb-8133-641566a114af {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 609.644656] env[61243]: DEBUG nova.scheduler.client.report [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.960360] env[61243]: DEBUG oslo_concurrency.lockutils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Releasing lock "refresh_cache-f21b7977-747e-437f-b03d-8442ed58756e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.960794] env[61243]: DEBUG nova.compute.manager [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 609.960972] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 609.961289] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-37913e4d-aa25-4b84-a391-81d9a724d07e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.970184] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3d01f7-5d94-4b33-a349-04cf74eb64c3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.994893] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f21b7977-747e-437f-b03d-8442ed58756e could not be found. [ 609.994995] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 609.995190] env[61243]: INFO nova.compute.manager [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 609.996211] env[61243]: DEBUG oslo.service.loopingcall [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 609.996211] env[61243]: DEBUG nova.compute.manager [-] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 609.996211] env[61243]: DEBUG nova.network.neutron [-] [instance: f21b7977-747e-437f-b03d-8442ed58756e] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 610.018742] env[61243]: DEBUG nova.network.neutron [-] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.095129] env[61243]: ERROR nova.compute.manager [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 432beef1-3f34-433c-9998-2c6210f10935, please check neutron logs for more information. [ 610.095129] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 610.095129] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.095129] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 610.095129] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.095129] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 610.095129] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.095129] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 610.095129] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.095129] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 610.095129] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.095129] env[61243]: ERROR nova.compute.manager raise self.value [ 610.095129] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.095129] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 610.095129] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.095129] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 610.095702] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.095702] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 610.095702] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 432beef1-3f34-433c-9998-2c6210f10935, please check neutron logs for more information. [ 610.095702] env[61243]: ERROR nova.compute.manager [ 610.095702] env[61243]: Traceback (most recent call last): [ 610.095702] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 610.095702] env[61243]: listener.cb(fileno) [ 610.095702] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.095702] env[61243]: result = function(*args, **kwargs) [ 610.095702] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.095702] env[61243]: return func(*args, **kwargs) [ 610.095702] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.095702] env[61243]: raise e [ 610.095702] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.095702] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 610.095702] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.095702] env[61243]: created_port_ids = self._update_ports_for_instance( [ 610.095702] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.095702] env[61243]: with excutils.save_and_reraise_exception(): [ 610.095702] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.095702] env[61243]: self.force_reraise() [ 610.095702] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.095702] env[61243]: raise self.value [ 610.095702] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.095702] env[61243]: updated_port = self._update_port( [ 610.095702] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.095702] env[61243]: _ensure_no_port_binding_failure(port) [ 610.095702] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.095702] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 610.096311] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 432beef1-3f34-433c-9998-2c6210f10935, please check neutron logs for more information. [ 610.096311] env[61243]: Removing descriptor: 16 [ 610.096311] env[61243]: ERROR nova.compute.manager [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 432beef1-3f34-433c-9998-2c6210f10935, please check neutron logs for more information. [ 610.096311] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Traceback (most recent call last): [ 610.096311] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 610.096311] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] yield resources [ 610.096311] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.096311] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] self.driver.spawn(context, instance, image_meta, [ 610.096311] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 610.096311] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.096311] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.096311] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] vm_ref = self.build_virtual_machine(instance, [ 610.096570] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.096570] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.096570] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.096570] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] for vif in network_info: [ 610.096570] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.096570] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] return self._sync_wrapper(fn, *args, **kwargs) [ 610.096570] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.096570] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] self.wait() [ 610.096570] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.096570] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] self[:] = self._gt.wait() [ 610.096570] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.096570] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] return self._exit_event.wait() [ 610.096570] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.096855] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] result = hub.switch() [ 610.096855] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.096855] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] return self.greenlet.switch() [ 610.096855] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.096855] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] result = function(*args, **kwargs) [ 610.096855] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.096855] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] return func(*args, **kwargs) [ 610.096855] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.096855] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] raise e [ 610.096855] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.096855] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] nwinfo = self.network_api.allocate_for_instance( [ 610.096855] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.096855] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] created_port_ids = self._update_ports_for_instance( [ 610.097141] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.097141] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] with excutils.save_and_reraise_exception(): [ 610.097141] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.097141] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] self.force_reraise() [ 610.097141] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.097141] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] raise self.value [ 610.097141] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.097141] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] updated_port = self._update_port( [ 610.097141] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.097141] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] _ensure_no_port_binding_failure(port) [ 610.097141] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.097141] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] raise exception.PortBindingFailed(port_id=port['id']) [ 610.097392] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] nova.exception.PortBindingFailed: Binding failed for port 432beef1-3f34-433c-9998-2c6210f10935, please check neutron logs for more information. [ 610.097392] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] [ 610.097392] env[61243]: INFO nova.compute.manager [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Terminating instance [ 610.098837] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Acquiring lock "refresh_cache-e2c498c5-3df1-4b99-a322-a598afd506f8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.098837] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Acquired lock "refresh_cache-e2c498c5-3df1-4b99-a322-a598afd506f8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.098837] env[61243]: DEBUG nova.network.neutron [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 610.150090] env[61243]: DEBUG oslo_concurrency.lockutils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.635s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.150090] env[61243]: DEBUG nova.compute.manager [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 610.152578] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.451s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.521150] env[61243]: DEBUG nova.network.neutron [-] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.619354] env[61243]: DEBUG nova.network.neutron [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.656669] env[61243]: DEBUG nova.compute.utils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 610.664901] env[61243]: DEBUG nova.compute.manager [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 610.665085] env[61243]: DEBUG nova.network.neutron [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 610.767784] env[61243]: DEBUG nova.policy [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1d64f20db125476a9e2cf5a77c88437e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f636baf6c50d49919d0c3286e83192b3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 610.799799] env[61243]: DEBUG nova.network.neutron [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.024512] env[61243]: INFO nova.compute.manager [-] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Took 1.03 seconds to deallocate network for instance. [ 611.028239] env[61243]: DEBUG nova.compute.claims [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 611.028239] env[61243]: DEBUG oslo_concurrency.lockutils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.169651] env[61243]: DEBUG nova.compute.manager [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 611.206247] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a057b29-a40a-4d35-99ba-088c1438821b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.218089] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b1ed4a-3342-48f2-8608-a37c6cd8a81e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.254339] env[61243]: DEBUG nova.network.neutron [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Successfully created port: 7cc00845-7608-4748-9fca-d4628e324770 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 611.256473] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-210b40bc-ab45-4e24-8d2d-b35ed66c87a4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.264902] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f5819fe-1bdf-4e3f-8633-43849a12d102 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.282161] env[61243]: DEBUG nova.compute.provider_tree [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.304249] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Releasing lock "refresh_cache-e2c498c5-3df1-4b99-a322-a598afd506f8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.304668] env[61243]: DEBUG nova.compute.manager [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 611.304859] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 611.305485] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-97124e9f-3fc8-40ab-922e-21c40dea7ee2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.314620] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c15fc7e-87ac-4af2-a239-7a2838102ec3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.337225] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e2c498c5-3df1-4b99-a322-a598afd506f8 could not be found. [ 611.337225] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 611.337225] env[61243]: INFO nova.compute.manager [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Took 0.03 seconds to destroy the instance on the hypervisor. [ 611.337225] env[61243]: DEBUG oslo.service.loopingcall [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 611.337465] env[61243]: DEBUG nova.compute.manager [-] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.337503] env[61243]: DEBUG nova.network.neutron [-] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 611.369019] env[61243]: DEBUG nova.network.neutron [-] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.670486] env[61243]: DEBUG nova.compute.manager [req-a52e2863-e96a-4621-a1f3-e0dd0ce0802f req-4f7e06ee-e656-4c34-b15f-c101b7c8db76 service nova] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Received event network-changed-432beef1-3f34-433c-9998-2c6210f10935 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 611.670751] env[61243]: DEBUG nova.compute.manager [req-a52e2863-e96a-4621-a1f3-e0dd0ce0802f req-4f7e06ee-e656-4c34-b15f-c101b7c8db76 service nova] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Refreshing instance network info cache due to event network-changed-432beef1-3f34-433c-9998-2c6210f10935. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 611.671205] env[61243]: DEBUG oslo_concurrency.lockutils [req-a52e2863-e96a-4621-a1f3-e0dd0ce0802f req-4f7e06ee-e656-4c34-b15f-c101b7c8db76 service nova] Acquiring lock "refresh_cache-e2c498c5-3df1-4b99-a322-a598afd506f8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.671390] env[61243]: DEBUG oslo_concurrency.lockutils [req-a52e2863-e96a-4621-a1f3-e0dd0ce0802f req-4f7e06ee-e656-4c34-b15f-c101b7c8db76 service nova] Acquired lock "refresh_cache-e2c498c5-3df1-4b99-a322-a598afd506f8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.671560] env[61243]: DEBUG nova.network.neutron [req-a52e2863-e96a-4621-a1f3-e0dd0ce0802f req-4f7e06ee-e656-4c34-b15f-c101b7c8db76 service nova] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Refreshing network info cache for port 432beef1-3f34-433c-9998-2c6210f10935 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 611.693558] env[61243]: DEBUG nova.network.neutron [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Successfully created port: ae6a1fb3-82f0-42ef-b406-3b4490e0a4d6 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 611.783138] env[61243]: DEBUG nova.scheduler.client.report [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.870526] env[61243]: DEBUG nova.network.neutron [-] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.137349] env[61243]: DEBUG nova.network.neutron [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Successfully created port: ebb338c0-06ef-45ad-a183-3deaf4af6ed1 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 612.180460] env[61243]: DEBUG nova.compute.manager [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 612.213689] env[61243]: DEBUG nova.virt.hardware [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 612.213689] env[61243]: DEBUG nova.virt.hardware [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 612.213689] env[61243]: DEBUG nova.virt.hardware [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 612.213689] env[61243]: DEBUG nova.virt.hardware [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 612.213907] env[61243]: DEBUG nova.virt.hardware [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 612.213907] env[61243]: DEBUG nova.virt.hardware [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 612.213907] env[61243]: DEBUG nova.virt.hardware [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 612.214126] env[61243]: DEBUG nova.virt.hardware [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 612.214334] env[61243]: DEBUG nova.virt.hardware [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 612.214551] env[61243]: DEBUG nova.virt.hardware [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 612.214770] env[61243]: DEBUG nova.virt.hardware [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 612.216292] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d4d19a2-5471-4414-a326-f3ec0227fdc6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.220191] env[61243]: DEBUG nova.network.neutron [req-a52e2863-e96a-4621-a1f3-e0dd0ce0802f req-4f7e06ee-e656-4c34-b15f-c101b7c8db76 service nova] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.226617] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e7a5b8-65eb-44b3-a182-d1ab9976d785 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.289642] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.137s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.290308] env[61243]: ERROR nova.compute.manager [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0ebbe785-0120-4fff-a6a9-8a38588de2ee, please check neutron logs for more information. [ 612.290308] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Traceback (most recent call last): [ 612.290308] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 612.290308] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] self.driver.spawn(context, instance, image_meta, [ 612.290308] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 612.290308] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 612.290308] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 612.290308] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] vm_ref = self.build_virtual_machine(instance, [ 612.290308] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 612.290308] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] vif_infos = vmwarevif.get_vif_info(self._session, [ 612.290308] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 612.290594] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] for vif in network_info: [ 612.290594] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 612.290594] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] return self._sync_wrapper(fn, *args, **kwargs) [ 612.290594] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 612.290594] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] self.wait() [ 612.290594] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 612.290594] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] self[:] = self._gt.wait() [ 612.290594] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 612.290594] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] return self._exit_event.wait() [ 612.290594] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 612.290594] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] current.throw(*self._exc) [ 612.290594] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.290594] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] result = function(*args, **kwargs) [ 612.290895] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 612.290895] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] return func(*args, **kwargs) [ 612.290895] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 612.290895] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] raise e [ 612.290895] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.290895] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] nwinfo = self.network_api.allocate_for_instance( [ 612.290895] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.290895] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] created_port_ids = self._update_ports_for_instance( [ 612.290895] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.290895] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] with excutils.save_and_reraise_exception(): [ 612.290895] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.290895] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] self.force_reraise() [ 612.290895] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.291206] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] raise self.value [ 612.291206] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.291206] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] updated_port = self._update_port( [ 612.291206] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.291206] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] _ensure_no_port_binding_failure(port) [ 612.291206] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.291206] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] raise exception.PortBindingFailed(port_id=port['id']) [ 612.291206] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] nova.exception.PortBindingFailed: Binding failed for port 0ebbe785-0120-4fff-a6a9-8a38588de2ee, please check neutron logs for more information. [ 612.291206] env[61243]: ERROR nova.compute.manager [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] [ 612.291206] env[61243]: DEBUG nova.compute.utils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Binding failed for port 0ebbe785-0120-4fff-a6a9-8a38588de2ee, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 612.296389] env[61243]: DEBUG nova.compute.manager [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Build of instance 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a was re-scheduled: Binding failed for port 0ebbe785-0120-4fff-a6a9-8a38588de2ee, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 612.296833] env[61243]: DEBUG nova.compute.manager [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 612.297081] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "refresh_cache-9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.297235] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "refresh_cache-9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.297396] env[61243]: DEBUG nova.network.neutron [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.298492] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.105s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.304410] env[61243]: INFO nova.compute.claims [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 612.373729] env[61243]: INFO nova.compute.manager [-] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Took 1.04 seconds to deallocate network for instance. [ 612.376710] env[61243]: DEBUG nova.compute.claims [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 612.376935] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.392190] env[61243]: DEBUG nova.network.neutron [req-a52e2863-e96a-4621-a1f3-e0dd0ce0802f req-4f7e06ee-e656-4c34-b15f-c101b7c8db76 service nova] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.835809] env[61243]: DEBUG nova.network.neutron [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.898351] env[61243]: DEBUG oslo_concurrency.lockutils [req-a52e2863-e96a-4621-a1f3-e0dd0ce0802f req-4f7e06ee-e656-4c34-b15f-c101b7c8db76 service nova] Releasing lock "refresh_cache-e2c498c5-3df1-4b99-a322-a598afd506f8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.898351] env[61243]: DEBUG nova.compute.manager [req-a52e2863-e96a-4621-a1f3-e0dd0ce0802f req-4f7e06ee-e656-4c34-b15f-c101b7c8db76 service nova] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Received event network-vif-deleted-432beef1-3f34-433c-9998-2c6210f10935 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 612.946940] env[61243]: DEBUG nova.network.neutron [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.370304] env[61243]: ERROR nova.compute.manager [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7cc00845-7608-4748-9fca-d4628e324770, please check neutron logs for more information. [ 613.370304] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 613.370304] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.370304] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 613.370304] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.370304] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 613.370304] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.370304] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 613.370304] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.370304] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 613.370304] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.370304] env[61243]: ERROR nova.compute.manager raise self.value [ 613.370304] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.370304] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 613.370304] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.370304] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 613.370982] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.370982] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 613.370982] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7cc00845-7608-4748-9fca-d4628e324770, please check neutron logs for more information. [ 613.370982] env[61243]: ERROR nova.compute.manager [ 613.370982] env[61243]: Traceback (most recent call last): [ 613.370982] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 613.370982] env[61243]: listener.cb(fileno) [ 613.370982] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.370982] env[61243]: result = function(*args, **kwargs) [ 613.370982] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 613.370982] env[61243]: return func(*args, **kwargs) [ 613.370982] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.370982] env[61243]: raise e [ 613.370982] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.370982] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 613.370982] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.370982] env[61243]: created_port_ids = self._update_ports_for_instance( [ 613.370982] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.370982] env[61243]: with excutils.save_and_reraise_exception(): [ 613.370982] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.370982] env[61243]: self.force_reraise() [ 613.370982] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.370982] env[61243]: raise self.value [ 613.370982] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.370982] env[61243]: updated_port = self._update_port( [ 613.370982] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.370982] env[61243]: _ensure_no_port_binding_failure(port) [ 613.370982] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.370982] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 613.371658] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 7cc00845-7608-4748-9fca-d4628e324770, please check neutron logs for more information. [ 613.371658] env[61243]: Removing descriptor: 15 [ 613.371658] env[61243]: ERROR nova.compute.manager [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7cc00845-7608-4748-9fca-d4628e324770, please check neutron logs for more information. [ 613.371658] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Traceback (most recent call last): [ 613.371658] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 613.371658] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] yield resources [ 613.371658] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.371658] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] self.driver.spawn(context, instance, image_meta, [ 613.371658] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 613.371658] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.371658] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.371658] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] vm_ref = self.build_virtual_machine(instance, [ 613.371985] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.371985] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.371985] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.371985] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] for vif in network_info: [ 613.371985] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.371985] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] return self._sync_wrapper(fn, *args, **kwargs) [ 613.371985] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.371985] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] self.wait() [ 613.371985] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.371985] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] self[:] = self._gt.wait() [ 613.371985] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.371985] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] return self._exit_event.wait() [ 613.371985] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.372276] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] result = hub.switch() [ 613.372276] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.372276] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] return self.greenlet.switch() [ 613.372276] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.372276] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] result = function(*args, **kwargs) [ 613.372276] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 613.372276] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] return func(*args, **kwargs) [ 613.372276] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.372276] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] raise e [ 613.372276] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.372276] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] nwinfo = self.network_api.allocate_for_instance( [ 613.372276] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.372276] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] created_port_ids = self._update_ports_for_instance( [ 613.372552] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.372552] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] with excutils.save_and_reraise_exception(): [ 613.372552] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.372552] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] self.force_reraise() [ 613.372552] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.372552] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] raise self.value [ 613.372552] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.372552] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] updated_port = self._update_port( [ 613.372552] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.372552] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] _ensure_no_port_binding_failure(port) [ 613.372552] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.372552] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] raise exception.PortBindingFailed(port_id=port['id']) [ 613.372806] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] nova.exception.PortBindingFailed: Binding failed for port 7cc00845-7608-4748-9fca-d4628e324770, please check neutron logs for more information. [ 613.372806] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] [ 613.372806] env[61243]: INFO nova.compute.manager [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Terminating instance [ 613.374782] env[61243]: DEBUG oslo_concurrency.lockutils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Acquiring lock "refresh_cache-5508a717-fb58-49b3-bfd1-b80d76a3c4e1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.374956] env[61243]: DEBUG oslo_concurrency.lockutils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Acquired lock "refresh_cache-5508a717-fb58-49b3-bfd1-b80d76a3c4e1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.375722] env[61243]: DEBUG nova.network.neutron [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 613.451823] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "refresh_cache-9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.452096] env[61243]: DEBUG nova.compute.manager [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 613.452254] env[61243]: DEBUG nova.compute.manager [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.452424] env[61243]: DEBUG nova.network.neutron [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.473430] env[61243]: DEBUG nova.network.neutron [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.613941] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Acquiring lock "2e25cb9a-9b21-4b60-b741-944e53bd50e9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.614101] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Lock "2e25cb9a-9b21-4b60-b741-944e53bd50e9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.810484] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bac7815-de39-462c-a29f-d8b51633765a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.818769] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0384e627-31a6-4720-be0b-5356cd076618 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.855876] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20dcb6e1-6813-40bb-b29b-c9b1926e36e7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.863691] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0942faef-2b2f-4aa8-afd2-223fa9d1e2fd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.878558] env[61243]: DEBUG nova.compute.provider_tree [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.915515] env[61243]: DEBUG nova.compute.manager [req-fc3b11cf-d946-4a6d-a70e-3f056538457b req-4899f1d5-a50a-48f3-a8e3-4be1cf2e18d4 service nova] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Received event network-changed-7cc00845-7608-4748-9fca-d4628e324770 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 613.915714] env[61243]: DEBUG nova.compute.manager [req-fc3b11cf-d946-4a6d-a70e-3f056538457b req-4899f1d5-a50a-48f3-a8e3-4be1cf2e18d4 service nova] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Refreshing instance network info cache due to event network-changed-7cc00845-7608-4748-9fca-d4628e324770. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 613.916075] env[61243]: DEBUG oslo_concurrency.lockutils [req-fc3b11cf-d946-4a6d-a70e-3f056538457b req-4899f1d5-a50a-48f3-a8e3-4be1cf2e18d4 service nova] Acquiring lock "refresh_cache-5508a717-fb58-49b3-bfd1-b80d76a3c4e1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.924449] env[61243]: DEBUG nova.network.neutron [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.977560] env[61243]: DEBUG nova.network.neutron [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.031479] env[61243]: DEBUG nova.network.neutron [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.381829] env[61243]: DEBUG nova.scheduler.client.report [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.483470] env[61243]: INFO nova.compute.manager [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a] Took 1.03 seconds to deallocate network for instance. [ 614.535422] env[61243]: DEBUG oslo_concurrency.lockutils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Releasing lock "refresh_cache-5508a717-fb58-49b3-bfd1-b80d76a3c4e1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.535854] env[61243]: DEBUG nova.compute.manager [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 614.536061] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 614.536368] env[61243]: DEBUG oslo_concurrency.lockutils [req-fc3b11cf-d946-4a6d-a70e-3f056538457b req-4899f1d5-a50a-48f3-a8e3-4be1cf2e18d4 service nova] Acquired lock "refresh_cache-5508a717-fb58-49b3-bfd1-b80d76a3c4e1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.536539] env[61243]: DEBUG nova.network.neutron [req-fc3b11cf-d946-4a6d-a70e-3f056538457b req-4899f1d5-a50a-48f3-a8e3-4be1cf2e18d4 service nova] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Refreshing network info cache for port 7cc00845-7608-4748-9fca-d4628e324770 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 614.537514] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3fbf5b70-faba-4b86-afc5-6b3f6d35d746 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.547300] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea1a09f-6c37-40ee-a439-e3f7105ad04f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.568240] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5508a717-fb58-49b3-bfd1-b80d76a3c4e1 could not be found. [ 614.568454] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 614.568631] env[61243]: INFO nova.compute.manager [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 614.568865] env[61243]: DEBUG oslo.service.loopingcall [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 614.569079] env[61243]: DEBUG nova.compute.manager [-] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.569179] env[61243]: DEBUG nova.network.neutron [-] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 614.633094] env[61243]: DEBUG nova.network.neutron [-] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.887076] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.588s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.887639] env[61243]: DEBUG nova.compute.manager [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 614.891963] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.374s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.055409] env[61243]: DEBUG nova.network.neutron [req-fc3b11cf-d946-4a6d-a70e-3f056538457b req-4899f1d5-a50a-48f3-a8e3-4be1cf2e18d4 service nova] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.132751] env[61243]: DEBUG nova.network.neutron [req-fc3b11cf-d946-4a6d-a70e-3f056538457b req-4899f1d5-a50a-48f3-a8e3-4be1cf2e18d4 service nova] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.396465] env[61243]: DEBUG nova.compute.utils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 615.400630] env[61243]: DEBUG nova.compute.manager [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 615.401233] env[61243]: DEBUG nova.network.neutron [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 615.441741] env[61243]: DEBUG nova.policy [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf35957b84954d238bd75756426033e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06c0bbb43a964f6fa145cfaa06ce35ac', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 615.502845] env[61243]: DEBUG nova.network.neutron [-] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.511602] env[61243]: INFO nova.scheduler.client.report [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Deleted allocations for instance 9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a [ 615.635184] env[61243]: DEBUG oslo_concurrency.lockutils [req-fc3b11cf-d946-4a6d-a70e-3f056538457b req-4899f1d5-a50a-48f3-a8e3-4be1cf2e18d4 service nova] Releasing lock "refresh_cache-5508a717-fb58-49b3-bfd1-b80d76a3c4e1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.635444] env[61243]: DEBUG nova.compute.manager [req-fc3b11cf-d946-4a6d-a70e-3f056538457b req-4899f1d5-a50a-48f3-a8e3-4be1cf2e18d4 service nova] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Received event network-vif-deleted-7cc00845-7608-4748-9fca-d4628e324770 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 615.703848] env[61243]: DEBUG nova.network.neutron [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Successfully created port: 538d5952-f284-4e64-99ae-1378fa45eb8a {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 615.854326] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5eb8fe1-fe9f-4e79-b9b8-f972bf22055b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.862911] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7806be35-ca00-440a-8fe4-fabe54efe8fb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.893023] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc69cd4-196d-40f0-955b-229c49facc72 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.899994] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee454c9-7627-4f0b-82b3-dc060ffc414b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.903933] env[61243]: DEBUG nova.compute.manager [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 615.916499] env[61243]: DEBUG nova.compute.provider_tree [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.006521] env[61243]: INFO nova.compute.manager [-] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Took 1.44 seconds to deallocate network for instance. [ 616.010093] env[61243]: DEBUG nova.compute.claims [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 616.010093] env[61243]: DEBUG oslo_concurrency.lockutils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.021892] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6f19c59-1373-4b6d-a385-b3673e84e9a8 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "9f2e0a06-b8c5-47e6-b8e3-d64fbf8fc47a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.313s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.323138] env[61243]: DEBUG nova.compute.manager [req-bde7c9ec-5c14-4cba-9a23-0beeb423e8a4 req-a55fa80e-33af-456a-b97a-bd499399682e service nova] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Received event network-changed-538d5952-f284-4e64-99ae-1378fa45eb8a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 616.323348] env[61243]: DEBUG nova.compute.manager [req-bde7c9ec-5c14-4cba-9a23-0beeb423e8a4 req-a55fa80e-33af-456a-b97a-bd499399682e service nova] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Refreshing instance network info cache due to event network-changed-538d5952-f284-4e64-99ae-1378fa45eb8a. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 616.323559] env[61243]: DEBUG oslo_concurrency.lockutils [req-bde7c9ec-5c14-4cba-9a23-0beeb423e8a4 req-a55fa80e-33af-456a-b97a-bd499399682e service nova] Acquiring lock "refresh_cache-476632bb-e07c-4660-9ea4-7290884e4761" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.323701] env[61243]: DEBUG oslo_concurrency.lockutils [req-bde7c9ec-5c14-4cba-9a23-0beeb423e8a4 req-a55fa80e-33af-456a-b97a-bd499399682e service nova] Acquired lock "refresh_cache-476632bb-e07c-4660-9ea4-7290884e4761" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.323858] env[61243]: DEBUG nova.network.neutron [req-bde7c9ec-5c14-4cba-9a23-0beeb423e8a4 req-a55fa80e-33af-456a-b97a-bd499399682e service nova] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Refreshing network info cache for port 538d5952-f284-4e64-99ae-1378fa45eb8a {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 616.422533] env[61243]: DEBUG nova.scheduler.client.report [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.510555] env[61243]: ERROR nova.compute.manager [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 538d5952-f284-4e64-99ae-1378fa45eb8a, please check neutron logs for more information. [ 616.510555] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 616.510555] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.510555] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 616.510555] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 616.510555] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 616.510555] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 616.510555] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 616.510555] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.510555] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 616.510555] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.510555] env[61243]: ERROR nova.compute.manager raise self.value [ 616.510555] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 616.510555] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 616.510555] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.510555] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 616.510922] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.510922] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 616.510922] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 538d5952-f284-4e64-99ae-1378fa45eb8a, please check neutron logs for more information. [ 616.510922] env[61243]: ERROR nova.compute.manager [ 616.510922] env[61243]: Traceback (most recent call last): [ 616.510922] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 616.510922] env[61243]: listener.cb(fileno) [ 616.510922] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.510922] env[61243]: result = function(*args, **kwargs) [ 616.510922] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 616.510922] env[61243]: return func(*args, **kwargs) [ 616.510922] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.510922] env[61243]: raise e [ 616.510922] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.510922] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 616.510922] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 616.510922] env[61243]: created_port_ids = self._update_ports_for_instance( [ 616.510922] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 616.510922] env[61243]: with excutils.save_and_reraise_exception(): [ 616.510922] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.510922] env[61243]: self.force_reraise() [ 616.510922] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.510922] env[61243]: raise self.value [ 616.510922] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 616.510922] env[61243]: updated_port = self._update_port( [ 616.510922] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.510922] env[61243]: _ensure_no_port_binding_failure(port) [ 616.510922] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.510922] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 616.511569] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 538d5952-f284-4e64-99ae-1378fa45eb8a, please check neutron logs for more information. [ 616.511569] env[61243]: Removing descriptor: 15 [ 616.525920] env[61243]: DEBUG nova.compute.manager [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 616.843581] env[61243]: DEBUG nova.network.neutron [req-bde7c9ec-5c14-4cba-9a23-0beeb423e8a4 req-a55fa80e-33af-456a-b97a-bd499399682e service nova] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.926995] env[61243]: DEBUG nova.network.neutron [req-bde7c9ec-5c14-4cba-9a23-0beeb423e8a4 req-a55fa80e-33af-456a-b97a-bd499399682e service nova] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.928905] env[61243]: DEBUG nova.compute.manager [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 616.931734] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.042s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.932327] env[61243]: ERROR nova.compute.manager [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d8ba96aa-f0b8-463f-bf1e-2e868a893c04, please check neutron logs for more information. [ 616.932327] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Traceback (most recent call last): [ 616.932327] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 616.932327] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] self.driver.spawn(context, instance, image_meta, [ 616.932327] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 616.932327] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.932327] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.932327] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] vm_ref = self.build_virtual_machine(instance, [ 616.932327] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.932327] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.932327] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.932659] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] for vif in network_info: [ 616.932659] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 616.932659] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] return self._sync_wrapper(fn, *args, **kwargs) [ 616.932659] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 616.932659] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] self.wait() [ 616.932659] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 616.932659] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] self[:] = self._gt.wait() [ 616.932659] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.932659] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] return self._exit_event.wait() [ 616.932659] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 616.932659] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] result = hub.switch() [ 616.932659] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 616.932659] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] return self.greenlet.switch() [ 616.932957] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.932957] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] result = function(*args, **kwargs) [ 616.932957] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 616.932957] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] return func(*args, **kwargs) [ 616.932957] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.932957] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] raise e [ 616.932957] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.932957] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] nwinfo = self.network_api.allocate_for_instance( [ 616.932957] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 616.932957] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] created_port_ids = self._update_ports_for_instance( [ 616.932957] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 616.932957] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] with excutils.save_and_reraise_exception(): [ 616.932957] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.933291] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] self.force_reraise() [ 616.933291] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.933291] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] raise self.value [ 616.933291] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 616.933291] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] updated_port = self._update_port( [ 616.933291] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.933291] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] _ensure_no_port_binding_failure(port) [ 616.933291] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.933291] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] raise exception.PortBindingFailed(port_id=port['id']) [ 616.933291] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] nova.exception.PortBindingFailed: Binding failed for port d8ba96aa-f0b8-463f-bf1e-2e868a893c04, please check neutron logs for more information. [ 616.933291] env[61243]: ERROR nova.compute.manager [instance: 762cbca7-4729-458a-b2af-2978723a25dc] [ 616.933570] env[61243]: DEBUG nova.compute.utils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Binding failed for port d8ba96aa-f0b8-463f-bf1e-2e868a893c04, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 616.934372] env[61243]: DEBUG nova.compute.manager [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Build of instance 762cbca7-4729-458a-b2af-2978723a25dc was re-scheduled: Binding failed for port d8ba96aa-f0b8-463f-bf1e-2e868a893c04, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 616.934844] env[61243]: DEBUG nova.compute.manager [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 616.935121] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Acquiring lock "refresh_cache-762cbca7-4729-458a-b2af-2978723a25dc" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.935288] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Acquired lock "refresh_cache-762cbca7-4729-458a-b2af-2978723a25dc" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.935450] env[61243]: DEBUG nova.network.neutron [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 616.937050] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.932s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.938819] env[61243]: INFO nova.compute.claims [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 616.959028] env[61243]: DEBUG nova.virt.hardware [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 616.959028] env[61243]: DEBUG nova.virt.hardware [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 616.959028] env[61243]: DEBUG nova.virt.hardware [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 616.959220] env[61243]: DEBUG nova.virt.hardware [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 616.959220] env[61243]: DEBUG nova.virt.hardware [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 616.959348] env[61243]: DEBUG nova.virt.hardware [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 616.959626] env[61243]: DEBUG nova.virt.hardware [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 616.959691] env[61243]: DEBUG nova.virt.hardware [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 616.959858] env[61243]: DEBUG nova.virt.hardware [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 616.960045] env[61243]: DEBUG nova.virt.hardware [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 616.960251] env[61243]: DEBUG nova.virt.hardware [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.961337] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd829b9-3b12-47b7-8895-a5889ec0e5ef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.970132] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81620504-595a-4550-8c9e-6566feb7369a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.986017] env[61243]: ERROR nova.compute.manager [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 538d5952-f284-4e64-99ae-1378fa45eb8a, please check neutron logs for more information. [ 616.986017] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Traceback (most recent call last): [ 616.986017] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 616.986017] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] yield resources [ 616.986017] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 616.986017] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] self.driver.spawn(context, instance, image_meta, [ 616.986017] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 616.986017] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.986017] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.986017] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] vm_ref = self.build_virtual_machine(instance, [ 616.986017] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.986356] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.986356] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.986356] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] for vif in network_info: [ 616.986356] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 616.986356] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] return self._sync_wrapper(fn, *args, **kwargs) [ 616.986356] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 616.986356] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] self.wait() [ 616.986356] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 616.986356] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] self[:] = self._gt.wait() [ 616.986356] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.986356] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] return self._exit_event.wait() [ 616.986356] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 616.986356] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] current.throw(*self._exc) [ 616.986637] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.986637] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] result = function(*args, **kwargs) [ 616.986637] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 616.986637] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] return func(*args, **kwargs) [ 616.986637] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.986637] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] raise e [ 616.986637] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.986637] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] nwinfo = self.network_api.allocate_for_instance( [ 616.986637] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 616.986637] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] created_port_ids = self._update_ports_for_instance( [ 616.986637] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 616.986637] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] with excutils.save_and_reraise_exception(): [ 616.986637] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.986932] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] self.force_reraise() [ 616.986932] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.986932] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] raise self.value [ 616.986932] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 616.986932] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] updated_port = self._update_port( [ 616.986932] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.986932] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] _ensure_no_port_binding_failure(port) [ 616.986932] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.986932] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] raise exception.PortBindingFailed(port_id=port['id']) [ 616.986932] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] nova.exception.PortBindingFailed: Binding failed for port 538d5952-f284-4e64-99ae-1378fa45eb8a, please check neutron logs for more information. [ 616.986932] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] [ 616.986932] env[61243]: INFO nova.compute.manager [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Terminating instance [ 616.988050] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquiring lock "refresh_cache-476632bb-e07c-4660-9ea4-7290884e4761" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.048297] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.440025] env[61243]: DEBUG oslo_concurrency.lockutils [req-bde7c9ec-5c14-4cba-9a23-0beeb423e8a4 req-a55fa80e-33af-456a-b97a-bd499399682e service nova] Releasing lock "refresh_cache-476632bb-e07c-4660-9ea4-7290884e4761" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.440476] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquired lock "refresh_cache-476632bb-e07c-4660-9ea4-7290884e4761" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.440664] env[61243]: DEBUG nova.network.neutron [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 617.456340] env[61243]: DEBUG nova.network.neutron [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.525878] env[61243]: DEBUG nova.network.neutron [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.763960] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "4a0a08ca-bbf5-4124-853e-0be45f531d91" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.764220] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "4a0a08ca-bbf5-4124-853e-0be45f531d91" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.963553] env[61243]: DEBUG nova.network.neutron [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.028399] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Releasing lock "refresh_cache-762cbca7-4729-458a-b2af-2978723a25dc" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.028399] env[61243]: DEBUG nova.compute.manager [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 618.028578] env[61243]: DEBUG nova.compute.manager [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.028787] env[61243]: DEBUG nova.network.neutron [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 618.046088] env[61243]: DEBUG nova.network.neutron [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.047736] env[61243]: DEBUG nova.network.neutron [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.324028] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6183c1f-4cc5-4b07-aabd-7d9e57976bb5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.333079] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267bab0a-c5fd-4350-81fe-95bb740a9b14 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.362483] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6f8e473-cfb8-4954-bd64-3b12ef3594b8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.370998] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b007fe10-40f9-46c9-98d9-a9fa155596e8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.384738] env[61243]: DEBUG nova.compute.provider_tree [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.394083] env[61243]: DEBUG nova.compute.manager [req-341a287c-5dde-40ae-a76a-b8e791c5b68f req-70128780-675b-4fa8-99a6-4a6140259bb6 service nova] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Received event network-vif-deleted-538d5952-f284-4e64-99ae-1378fa45eb8a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 618.550420] env[61243]: DEBUG nova.network.neutron [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.551772] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Releasing lock "refresh_cache-476632bb-e07c-4660-9ea4-7290884e4761" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.552169] env[61243]: DEBUG nova.compute.manager [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 618.552383] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 618.552837] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd608d08-2dea-4899-b1c9-de572c4e7fc1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.561592] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be11dcb7-d100-427e-85b9-b51d81e61f47 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.585876] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 476632bb-e07c-4660-9ea4-7290884e4761 could not be found. [ 618.586081] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 618.586263] env[61243]: INFO nova.compute.manager [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Took 0.03 seconds to destroy the instance on the hypervisor. [ 618.586504] env[61243]: DEBUG oslo.service.loopingcall [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 618.586709] env[61243]: DEBUG nova.compute.manager [-] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.586802] env[61243]: DEBUG nova.network.neutron [-] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 618.601849] env[61243]: DEBUG nova.network.neutron [-] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.888667] env[61243]: DEBUG nova.scheduler.client.report [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 619.054332] env[61243]: INFO nova.compute.manager [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] [instance: 762cbca7-4729-458a-b2af-2978723a25dc] Took 1.03 seconds to deallocate network for instance. [ 619.104055] env[61243]: DEBUG nova.network.neutron [-] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.393487] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.394969] env[61243]: DEBUG nova.compute.manager [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 619.396556] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.507s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.398012] env[61243]: INFO nova.compute.claims [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.606631] env[61243]: INFO nova.compute.manager [-] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Took 1.02 seconds to deallocate network for instance. [ 619.609070] env[61243]: DEBUG nova.compute.claims [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 619.609363] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.903058] env[61243]: DEBUG nova.compute.utils [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 619.906093] env[61243]: DEBUG nova.compute.manager [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Not allocating networking since 'none' was specified. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 620.084943] env[61243]: INFO nova.scheduler.client.report [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Deleted allocations for instance 762cbca7-4729-458a-b2af-2978723a25dc [ 620.407334] env[61243]: DEBUG nova.compute.manager [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 620.592493] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2e66fc9a-8c34-4a40-bf05-ed529a149088 tempest-ServersWithSpecificFlavorTestJSON-569636144 tempest-ServersWithSpecificFlavorTestJSON-569636144-project-member] Lock "762cbca7-4729-458a-b2af-2978723a25dc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.541s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.784067] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c4b1edd-fed0-41f5-b55c-475d08d54e17 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.792036] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc401708-291e-4428-84df-89f103187254 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.417300] env[61243]: DEBUG nova.compute.manager [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 621.420431] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36b518c-4266-4d84-bee0-f18375be7415 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.430248] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2517db2a-5d42-407b-a9a1-a37b1a441592 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.446239] env[61243]: DEBUG nova.compute.provider_tree [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.922089] env[61243]: DEBUG nova.compute.manager [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 621.945183] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.949623] env[61243]: DEBUG nova.scheduler.client.report [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.963140] env[61243]: DEBUG nova.virt.hardware [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 621.963393] env[61243]: DEBUG nova.virt.hardware [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 621.963549] env[61243]: DEBUG nova.virt.hardware [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 621.963729] env[61243]: DEBUG nova.virt.hardware [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 621.963906] env[61243]: DEBUG nova.virt.hardware [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 621.964569] env[61243]: DEBUG nova.virt.hardware [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 621.964852] env[61243]: DEBUG nova.virt.hardware [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 621.965044] env[61243]: DEBUG nova.virt.hardware [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 621.965223] env[61243]: DEBUG nova.virt.hardware [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 621.965428] env[61243]: DEBUG nova.virt.hardware [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 621.965681] env[61243]: DEBUG nova.virt.hardware [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 621.966676] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d0e9ac-263f-4222-a00b-abc6bce72854 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.976027] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf56fd53-28db-420f-bf87-19a3b9fbd689 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.988782] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Instance VIF info [] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 621.996169] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Creating folder: Project (2c5d8b94831c4f0ebc101e9c225cd139). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 621.996462] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c4024198-9ff9-4747-896a-11a3442d3e80 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.006832] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Created folder: Project (2c5d8b94831c4f0ebc101e9c225cd139) in parent group-v285636. [ 622.007028] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Creating folder: Instances. Parent ref: group-v285650. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 622.007252] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a4035e27-b73d-494d-a859-254bab0d29db {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.016010] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Created folder: Instances in parent group-v285650. [ 622.016681] env[61243]: DEBUG oslo.service.loopingcall [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 622.016681] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 622.016681] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7b2f0d3-b0c9-4f35-a130-ef269ac35ad7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.031983] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 622.031983] env[61243]: value = "task-1338738" [ 622.031983] env[61243]: _type = "Task" [ 622.031983] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.045896] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338738, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.454393] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.058s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.454915] env[61243]: DEBUG nova.compute.manager [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 622.457540] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.407s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.541494] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338738, 'name': CreateVM_Task, 'duration_secs': 0.259976} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.541709] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 622.542084] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.542237] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.542550] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 622.542791] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4fad8e2-2be0-4689-bdef-698c3239d565 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.547923] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Waiting for the task: (returnval){ [ 622.547923] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523a9a4f-fc51-856b-b003-98f32448984d" [ 622.547923] env[61243]: _type = "Task" [ 622.547923] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.555378] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523a9a4f-fc51-856b-b003-98f32448984d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.963585] env[61243]: DEBUG nova.compute.utils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 622.967810] env[61243]: DEBUG nova.compute.manager [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 622.968325] env[61243]: DEBUG nova.network.neutron [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 623.023822] env[61243]: DEBUG nova.policy [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf35957b84954d238bd75756426033e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06c0bbb43a964f6fa145cfaa06ce35ac', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 623.060211] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523a9a4f-fc51-856b-b003-98f32448984d, 'name': SearchDatastore_Task, 'duration_secs': 0.014107} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.060577] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.060838] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 623.061119] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.061328] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.061528] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 623.061815] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-73b44da0-3253-44f4-aa10-c0dde2975988 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.074491] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 623.075141] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 623.075589] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0815ca5f-430b-458d-81f6-147a6a782f0c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.083316] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Waiting for the task: (returnval){ [ 623.083316] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5238f3ea-a18b-1c4a-fdc8-02217e92c26f" [ 623.083316] env[61243]: _type = "Task" [ 623.083316] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.091669] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5238f3ea-a18b-1c4a-fdc8-02217e92c26f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.314576] env[61243]: DEBUG nova.network.neutron [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Successfully created port: 740571e3-5361-4bd1-abce-b7699711237f {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 623.442119] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-481d51ed-d72f-43c0-8829-3f5f998cbf9c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.450049] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92af120c-8b77-478d-8abe-177342c24c2e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.479670] env[61243]: DEBUG nova.compute.manager [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 623.483374] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25572d36-4d70-4dd5-b95a-44f602eec609 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.491500] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9c85f8-01d6-4fa8-b542-fcdf3d1a2cd4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.505868] env[61243]: DEBUG nova.compute.provider_tree [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.594053] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5238f3ea-a18b-1c4a-fdc8-02217e92c26f, 'name': SearchDatastore_Task, 'duration_secs': 0.008422} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.594910] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ce1372a-6bcd-4a40-bf48-d65bfd9b9227 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.599949] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Waiting for the task: (returnval){ [ 623.599949] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52252334-5209-8d4b-2401-59ae71b45966" [ 623.599949] env[61243]: _type = "Task" [ 623.599949] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.607590] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52252334-5209-8d4b-2401-59ae71b45966, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.008791] env[61243]: DEBUG nova.scheduler.client.report [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.117314] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52252334-5209-8d4b-2401-59ae71b45966, 'name': SearchDatastore_Task, 'duration_secs': 0.008592} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.117627] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.118122] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] b5163d89-ceb8-4c61-ae02-0ae3311b58a9/b5163d89-ceb8-4c61-ae02-0ae3311b58a9.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 624.118439] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-daf9fb3e-f224-47df-ad20-7a553a6c83c3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.125176] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Waiting for the task: (returnval){ [ 624.125176] env[61243]: value = "task-1338739" [ 624.125176] env[61243]: _type = "Task" [ 624.125176] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.134651] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': task-1338739, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.194718] env[61243]: DEBUG nova.compute.manager [req-4fb14ff4-1574-488e-a711-98e07b90794a req-a9a7b434-e459-47af-89aa-e4ef5964e6e8 service nova] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Received event network-changed-740571e3-5361-4bd1-abce-b7699711237f {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 624.194956] env[61243]: DEBUG nova.compute.manager [req-4fb14ff4-1574-488e-a711-98e07b90794a req-a9a7b434-e459-47af-89aa-e4ef5964e6e8 service nova] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Refreshing instance network info cache due to event network-changed-740571e3-5361-4bd1-abce-b7699711237f. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 624.195284] env[61243]: DEBUG oslo_concurrency.lockutils [req-4fb14ff4-1574-488e-a711-98e07b90794a req-a9a7b434-e459-47af-89aa-e4ef5964e6e8 service nova] Acquiring lock "refresh_cache-461b2ef2-d665-47ab-affa-e525bffe561a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.195464] env[61243]: DEBUG oslo_concurrency.lockutils [req-4fb14ff4-1574-488e-a711-98e07b90794a req-a9a7b434-e459-47af-89aa-e4ef5964e6e8 service nova] Acquired lock "refresh_cache-461b2ef2-d665-47ab-affa-e525bffe561a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.195658] env[61243]: DEBUG nova.network.neutron [req-4fb14ff4-1574-488e-a711-98e07b90794a req-a9a7b434-e459-47af-89aa-e4ef5964e6e8 service nova] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Refreshing network info cache for port 740571e3-5361-4bd1-abce-b7699711237f {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 624.319661] env[61243]: ERROR nova.compute.manager [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 740571e3-5361-4bd1-abce-b7699711237f, please check neutron logs for more information. [ 624.319661] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 624.319661] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.319661] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 624.319661] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.319661] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 624.319661] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.319661] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 624.319661] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.319661] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 624.319661] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.319661] env[61243]: ERROR nova.compute.manager raise self.value [ 624.319661] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.319661] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 624.319661] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.319661] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 624.320127] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.320127] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 624.320127] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 740571e3-5361-4bd1-abce-b7699711237f, please check neutron logs for more information. [ 624.320127] env[61243]: ERROR nova.compute.manager [ 624.320127] env[61243]: Traceback (most recent call last): [ 624.320127] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 624.320127] env[61243]: listener.cb(fileno) [ 624.320127] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.320127] env[61243]: result = function(*args, **kwargs) [ 624.320127] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 624.320127] env[61243]: return func(*args, **kwargs) [ 624.320127] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.320127] env[61243]: raise e [ 624.320127] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.320127] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 624.320127] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.320127] env[61243]: created_port_ids = self._update_ports_for_instance( [ 624.320127] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.320127] env[61243]: with excutils.save_and_reraise_exception(): [ 624.320127] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.320127] env[61243]: self.force_reraise() [ 624.320127] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.320127] env[61243]: raise self.value [ 624.320127] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.320127] env[61243]: updated_port = self._update_port( [ 624.320127] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.320127] env[61243]: _ensure_no_port_binding_failure(port) [ 624.320127] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.320127] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 624.320880] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 740571e3-5361-4bd1-abce-b7699711237f, please check neutron logs for more information. [ 624.320880] env[61243]: Removing descriptor: 16 [ 624.492706] env[61243]: DEBUG nova.compute.manager [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 624.514768] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.057s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.515896] env[61243]: ERROR nova.compute.manager [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 92c79f78-3449-4425-a9d2-a2df126769b9, please check neutron logs for more information. [ 624.515896] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] Traceback (most recent call last): [ 624.515896] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.515896] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] self.driver.spawn(context, instance, image_meta, [ 624.515896] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 624.515896] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.515896] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.515896] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] vm_ref = self.build_virtual_machine(instance, [ 624.515896] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.515896] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.515896] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.517163] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] for vif in network_info: [ 624.517163] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.517163] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] return self._sync_wrapper(fn, *args, **kwargs) [ 624.517163] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.517163] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] self.wait() [ 624.517163] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.517163] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] self[:] = self._gt.wait() [ 624.517163] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.517163] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] return self._exit_event.wait() [ 624.517163] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.517163] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] result = hub.switch() [ 624.517163] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.517163] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] return self.greenlet.switch() [ 624.519120] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.519120] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] result = function(*args, **kwargs) [ 624.519120] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 624.519120] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] return func(*args, **kwargs) [ 624.519120] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.519120] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] raise e [ 624.519120] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.519120] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] nwinfo = self.network_api.allocate_for_instance( [ 624.519120] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.519120] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] created_port_ids = self._update_ports_for_instance( [ 624.519120] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.519120] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] with excutils.save_and_reraise_exception(): [ 624.519120] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.519691] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] self.force_reraise() [ 624.519691] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.519691] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] raise self.value [ 624.519691] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.519691] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] updated_port = self._update_port( [ 624.519691] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.519691] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] _ensure_no_port_binding_failure(port) [ 624.519691] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.519691] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] raise exception.PortBindingFailed(port_id=port['id']) [ 624.519691] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] nova.exception.PortBindingFailed: Binding failed for port 92c79f78-3449-4425-a9d2-a2df126769b9, please check neutron logs for more information. [ 624.519691] env[61243]: ERROR nova.compute.manager [instance: 3700deef-a438-4920-b35f-8fba370a5984] [ 624.519943] env[61243]: DEBUG nova.compute.utils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Binding failed for port 92c79f78-3449-4425-a9d2-a2df126769b9, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 624.519943] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.670s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.522507] env[61243]: DEBUG nova.compute.manager [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Build of instance 3700deef-a438-4920-b35f-8fba370a5984 was re-scheduled: Binding failed for port 92c79f78-3449-4425-a9d2-a2df126769b9, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 624.522507] env[61243]: DEBUG nova.compute.manager [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 624.522507] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Acquiring lock "refresh_cache-3700deef-a438-4920-b35f-8fba370a5984" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.522723] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Acquired lock "refresh_cache-3700deef-a438-4920-b35f-8fba370a5984" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.522789] env[61243]: DEBUG nova.network.neutron [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 624.526264] env[61243]: DEBUG nova.virt.hardware [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 624.526510] env[61243]: DEBUG nova.virt.hardware [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 624.526736] env[61243]: DEBUG nova.virt.hardware [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 624.526981] env[61243]: DEBUG nova.virt.hardware [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 624.529547] env[61243]: DEBUG nova.virt.hardware [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 624.529800] env[61243]: DEBUG nova.virt.hardware [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 624.530111] env[61243]: DEBUG nova.virt.hardware [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 624.530309] env[61243]: DEBUG nova.virt.hardware [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 624.530566] env[61243]: DEBUG nova.virt.hardware [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 624.530929] env[61243]: DEBUG nova.virt.hardware [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 624.531260] env[61243]: DEBUG nova.virt.hardware [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 624.532470] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab4d19a-5624-415f-b1b4-0db8b5f90d1f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.551829] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc80a4f-4ec0-4f09-bf48-620f9ca1e245 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.571603] env[61243]: ERROR nova.compute.manager [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 740571e3-5361-4bd1-abce-b7699711237f, please check neutron logs for more information. [ 624.571603] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Traceback (most recent call last): [ 624.571603] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 624.571603] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] yield resources [ 624.571603] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.571603] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] self.driver.spawn(context, instance, image_meta, [ 624.571603] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 624.571603] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.571603] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.571603] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] vm_ref = self.build_virtual_machine(instance, [ 624.571603] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.571977] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.571977] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.571977] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] for vif in network_info: [ 624.571977] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.571977] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] return self._sync_wrapper(fn, *args, **kwargs) [ 624.571977] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.571977] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] self.wait() [ 624.571977] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.571977] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] self[:] = self._gt.wait() [ 624.571977] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.571977] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] return self._exit_event.wait() [ 624.571977] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 624.571977] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] current.throw(*self._exc) [ 624.572357] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.572357] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] result = function(*args, **kwargs) [ 624.572357] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 624.572357] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] return func(*args, **kwargs) [ 624.572357] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.572357] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] raise e [ 624.572357] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.572357] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] nwinfo = self.network_api.allocate_for_instance( [ 624.572357] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.572357] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] created_port_ids = self._update_ports_for_instance( [ 624.572357] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.572357] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] with excutils.save_and_reraise_exception(): [ 624.572357] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.572706] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] self.force_reraise() [ 624.572706] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.572706] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] raise self.value [ 624.572706] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.572706] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] updated_port = self._update_port( [ 624.572706] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.572706] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] _ensure_no_port_binding_failure(port) [ 624.572706] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.572706] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] raise exception.PortBindingFailed(port_id=port['id']) [ 624.572706] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] nova.exception.PortBindingFailed: Binding failed for port 740571e3-5361-4bd1-abce-b7699711237f, please check neutron logs for more information. [ 624.572706] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] [ 624.572706] env[61243]: INFO nova.compute.manager [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Terminating instance [ 624.574419] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquiring lock "refresh_cache-461b2ef2-d665-47ab-affa-e525bffe561a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.637479] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': task-1338739, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471335} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.637749] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] b5163d89-ceb8-4c61-ae02-0ae3311b58a9/b5163d89-ceb8-4c61-ae02-0ae3311b58a9.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 624.637967] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 624.638519] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1a054905-fbe9-4251-8861-3ee124a4eaef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.644958] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Waiting for the task: (returnval){ [ 624.644958] env[61243]: value = "task-1338740" [ 624.644958] env[61243]: _type = "Task" [ 624.644958] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.652427] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': task-1338740, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.715584] env[61243]: DEBUG nova.network.neutron [req-4fb14ff4-1574-488e-a711-98e07b90794a req-a9a7b434-e459-47af-89aa-e4ef5964e6e8 service nova] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.795963] env[61243]: DEBUG nova.network.neutron [req-4fb14ff4-1574-488e-a711-98e07b90794a req-a9a7b434-e459-47af-89aa-e4ef5964e6e8 service nova] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.061640] env[61243]: DEBUG nova.network.neutron [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.156832] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': task-1338740, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064988} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.157114] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 625.157877] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32d3174-49ab-47fc-95d4-291f82996fbb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.179397] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Reconfiguring VM instance instance-00000014 to attach disk [datastore2] b5163d89-ceb8-4c61-ae02-0ae3311b58a9/b5163d89-ceb8-4c61-ae02-0ae3311b58a9.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 625.184015] env[61243]: DEBUG nova.network.neutron [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.184015] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6912d03-5d58-4725-a56d-2ef08b4aada2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.199795] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Releasing lock "refresh_cache-3700deef-a438-4920-b35f-8fba370a5984" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.200023] env[61243]: DEBUG nova.compute.manager [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 625.200230] env[61243]: DEBUG nova.compute.manager [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.203058] env[61243]: DEBUG nova.network.neutron [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 625.207516] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Waiting for the task: (returnval){ [ 625.207516] env[61243]: value = "task-1338741" [ 625.207516] env[61243]: _type = "Task" [ 625.207516] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.219018] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': task-1338741, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.219018] env[61243]: DEBUG nova.network.neutron [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.298991] env[61243]: DEBUG oslo_concurrency.lockutils [req-4fb14ff4-1574-488e-a711-98e07b90794a req-a9a7b434-e459-47af-89aa-e4ef5964e6e8 service nova] Releasing lock "refresh_cache-461b2ef2-d665-47ab-affa-e525bffe561a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.299757] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquired lock "refresh_cache-461b2ef2-d665-47ab-affa-e525bffe561a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.300119] env[61243]: DEBUG nova.network.neutron [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 625.603070] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-785d5c0a-a8ec-4190-ba6f-752e11d95c5c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.610331] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff7c51a1-cf9e-4fbf-a51b-f357cf1dbccb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.640508] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df2414bb-f4dd-46c7-bdbc-62a5cc0bf51f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.649036] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aea6c97-f9ac-47ea-bc80-573c27bf5ab5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.660669] env[61243]: DEBUG nova.compute.provider_tree [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.717282] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': task-1338741, 'name': ReconfigVM_Task, 'duration_secs': 0.308355} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.717556] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Reconfigured VM instance instance-00000014 to attach disk [datastore2] b5163d89-ceb8-4c61-ae02-0ae3311b58a9/b5163d89-ceb8-4c61-ae02-0ae3311b58a9.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 625.718199] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8db8e9ed-a14f-4f61-b097-f7bdde8b0707 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.719848] env[61243]: DEBUG nova.network.neutron [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.725340] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Waiting for the task: (returnval){ [ 625.725340] env[61243]: value = "task-1338742" [ 625.725340] env[61243]: _type = "Task" [ 625.725340] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.734203] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': task-1338742, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.820098] env[61243]: DEBUG nova.network.neutron [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.905735] env[61243]: DEBUG nova.network.neutron [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.165669] env[61243]: DEBUG nova.scheduler.client.report [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.216678] env[61243]: DEBUG nova.compute.manager [req-a5c1ed3f-1a86-4577-b448-6cf94eeb48ae req-555870e9-fdd6-4558-b6a3-5d3fbc08079f service nova] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Received event network-vif-deleted-740571e3-5361-4bd1-abce-b7699711237f {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 626.222728] env[61243]: INFO nova.compute.manager [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] [instance: 3700deef-a438-4920-b35f-8fba370a5984] Took 1.02 seconds to deallocate network for instance. [ 626.235329] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': task-1338742, 'name': Rename_Task, 'duration_secs': 0.147626} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.235464] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 626.235706] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-399f6a1b-acfc-4ac0-9bad-d171bc8a607f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.242352] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Waiting for the task: (returnval){ [ 626.242352] env[61243]: value = "task-1338743" [ 626.242352] env[61243]: _type = "Task" [ 626.242352] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.250132] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': task-1338743, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.409079] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Releasing lock "refresh_cache-461b2ef2-d665-47ab-affa-e525bffe561a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.409595] env[61243]: DEBUG nova.compute.manager [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 626.410469] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 626.410469] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a60c6f84-c83c-4030-ba59-be8c665af99e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.418295] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7acac3b7-8835-44fa-989b-2e5f54d6c603 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.441182] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 461b2ef2-d665-47ab-affa-e525bffe561a could not be found. [ 626.441425] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 626.441602] env[61243]: INFO nova.compute.manager [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 626.441837] env[61243]: DEBUG oslo.service.loopingcall [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 626.442059] env[61243]: DEBUG nova.compute.manager [-] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.442158] env[61243]: DEBUG nova.network.neutron [-] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 626.457409] env[61243]: DEBUG nova.network.neutron [-] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.671314] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.153s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.671935] env[61243]: ERROR nova.compute.manager [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 74fc99ca-72cf-4729-bf32-21f722cbdaf0, please check neutron logs for more information. [ 626.671935] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] Traceback (most recent call last): [ 626.671935] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 626.671935] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] self.driver.spawn(context, instance, image_meta, [ 626.671935] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 626.671935] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.671935] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.671935] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] vm_ref = self.build_virtual_machine(instance, [ 626.671935] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.671935] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.671935] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.672238] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] for vif in network_info: [ 626.672238] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 626.672238] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] return self._sync_wrapper(fn, *args, **kwargs) [ 626.672238] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 626.672238] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] self.wait() [ 626.672238] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 626.672238] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] self[:] = self._gt.wait() [ 626.672238] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.672238] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] return self._exit_event.wait() [ 626.672238] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 626.672238] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] current.throw(*self._exc) [ 626.672238] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.672238] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] result = function(*args, **kwargs) [ 626.672551] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.672551] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] return func(*args, **kwargs) [ 626.672551] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.672551] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] raise e [ 626.672551] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.672551] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] nwinfo = self.network_api.allocate_for_instance( [ 626.672551] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.672551] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] created_port_ids = self._update_ports_for_instance( [ 626.672551] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.672551] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] with excutils.save_and_reraise_exception(): [ 626.672551] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.672551] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] self.force_reraise() [ 626.672551] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.672981] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] raise self.value [ 626.672981] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.672981] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] updated_port = self._update_port( [ 626.672981] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.672981] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] _ensure_no_port_binding_failure(port) [ 626.672981] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.672981] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] raise exception.PortBindingFailed(port_id=port['id']) [ 626.672981] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] nova.exception.PortBindingFailed: Binding failed for port 74fc99ca-72cf-4729-bf32-21f722cbdaf0, please check neutron logs for more information. [ 626.672981] env[61243]: ERROR nova.compute.manager [instance: 2451843a-fd32-450e-933e-e635c789834b] [ 626.672981] env[61243]: DEBUG nova.compute.utils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Binding failed for port 74fc99ca-72cf-4729-bf32-21f722cbdaf0, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 626.674745] env[61243]: DEBUG oslo_concurrency.lockutils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.646s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.677869] env[61243]: DEBUG nova.compute.manager [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Build of instance 2451843a-fd32-450e-933e-e635c789834b was re-scheduled: Binding failed for port 74fc99ca-72cf-4729-bf32-21f722cbdaf0, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 626.678438] env[61243]: DEBUG nova.compute.manager [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 626.678553] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Acquiring lock "refresh_cache-2451843a-fd32-450e-933e-e635c789834b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.678680] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Acquired lock "refresh_cache-2451843a-fd32-450e-933e-e635c789834b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.678840] env[61243]: DEBUG nova.network.neutron [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 626.752676] env[61243]: DEBUG oslo_vmware.api [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': task-1338743, 'name': PowerOnVM_Task, 'duration_secs': 0.402027} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.753026] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 626.753142] env[61243]: INFO nova.compute.manager [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Took 4.83 seconds to spawn the instance on the hypervisor. [ 626.753344] env[61243]: DEBUG nova.compute.manager [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 626.754033] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0256bc0b-9c93-4fb6-a2f2-07f1c1772320 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.961319] env[61243]: DEBUG nova.network.neutron [-] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.200326] env[61243]: DEBUG nova.network.neutron [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.262509] env[61243]: INFO nova.scheduler.client.report [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Deleted allocations for instance 3700deef-a438-4920-b35f-8fba370a5984 [ 627.278884] env[61243]: INFO nova.compute.manager [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Took 28.30 seconds to build instance. [ 627.318517] env[61243]: DEBUG nova.network.neutron [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.447362] env[61243]: DEBUG nova.compute.manager [None req-8688da4e-8a65-4cc1-b82d-91116d282e4c tempest-ServerDiagnosticsV248Test-942551326 tempest-ServerDiagnosticsV248Test-942551326-project-admin] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 627.448489] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f09358a-7e8a-4d5e-a46c-dab54aa0eedb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.457483] env[61243]: INFO nova.compute.manager [None req-8688da4e-8a65-4cc1-b82d-91116d282e4c tempest-ServerDiagnosticsV248Test-942551326 tempest-ServerDiagnosticsV248Test-942551326-project-admin] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Retrieving diagnostics [ 627.458858] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67f7c9f-4184-4082-bc01-9b829e823f87 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.462574] env[61243]: INFO nova.compute.manager [-] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Took 1.02 seconds to deallocate network for instance. [ 627.493819] env[61243]: DEBUG nova.compute.claims [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 627.494029] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.729259] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73bfed1-16dd-4895-9440-1d375483a406 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.739636] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d92742-da0f-4801-a9f6-31358ff8f874 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.768221] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5282ece7-6cbd-4421-89e7-e3e50467ec6c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.774815] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f942807-829a-426b-9425-b65df129ada4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.778833] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46f4ca99-1c00-49c3-b43a-f8e48d95cdc2 tempest-ServersV294TestFqdnHostnames-1551344508 tempest-ServersV294TestFqdnHostnames-1551344508-project-member] Lock "3700deef-a438-4920-b35f-8fba370a5984" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.156s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.781035] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0ca8671b-4d1e-4b4a-a860-3f8dc8f0a5a4 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Lock "b5163d89-ceb8-4c61-ae02-0ae3311b58a9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.621s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.789785] env[61243]: DEBUG nova.compute.provider_tree [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.792176] env[61243]: DEBUG nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 627.821437] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Releasing lock "refresh_cache-2451843a-fd32-450e-933e-e635c789834b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.821972] env[61243]: DEBUG nova.compute.manager [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 627.821972] env[61243]: DEBUG nova.compute.manager [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 627.822094] env[61243]: DEBUG nova.network.neutron [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 627.838400] env[61243]: DEBUG nova.network.neutron [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.281852] env[61243]: DEBUG nova.compute.manager [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 628.296744] env[61243]: DEBUG nova.scheduler.client.report [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.319741] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.341237] env[61243]: DEBUG nova.network.neutron [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.801817] env[61243]: DEBUG oslo_concurrency.lockutils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.127s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.802498] env[61243]: ERROR nova.compute.manager [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a5db5193-898e-47cb-8133-641566a114af, please check neutron logs for more information. [ 628.802498] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] Traceback (most recent call last): [ 628.802498] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 628.802498] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] self.driver.spawn(context, instance, image_meta, [ 628.802498] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 628.802498] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.802498] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.802498] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] vm_ref = self.build_virtual_machine(instance, [ 628.802498] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.802498] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.802498] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.802818] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] for vif in network_info: [ 628.802818] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 628.802818] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] return self._sync_wrapper(fn, *args, **kwargs) [ 628.802818] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 628.802818] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] self.wait() [ 628.802818] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 628.802818] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] self[:] = self._gt.wait() [ 628.802818] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.802818] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] return self._exit_event.wait() [ 628.802818] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.802818] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] result = hub.switch() [ 628.802818] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.802818] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] return self.greenlet.switch() [ 628.803327] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.803327] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] result = function(*args, **kwargs) [ 628.803327] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 628.803327] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] return func(*args, **kwargs) [ 628.803327] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.803327] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] raise e [ 628.803327] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.803327] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] nwinfo = self.network_api.allocate_for_instance( [ 628.803327] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.803327] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] created_port_ids = self._update_ports_for_instance( [ 628.803327] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.803327] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] with excutils.save_and_reraise_exception(): [ 628.803327] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.803622] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] self.force_reraise() [ 628.803622] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.803622] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] raise self.value [ 628.803622] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.803622] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] updated_port = self._update_port( [ 628.803622] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.803622] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] _ensure_no_port_binding_failure(port) [ 628.803622] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.803622] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] raise exception.PortBindingFailed(port_id=port['id']) [ 628.803622] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] nova.exception.PortBindingFailed: Binding failed for port a5db5193-898e-47cb-8133-641566a114af, please check neutron logs for more information. [ 628.803622] env[61243]: ERROR nova.compute.manager [instance: f21b7977-747e-437f-b03d-8442ed58756e] [ 628.803871] env[61243]: DEBUG nova.compute.utils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Binding failed for port a5db5193-898e-47cb-8133-641566a114af, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 628.805527] env[61243]: DEBUG oslo_concurrency.lockutils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.805684] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.429s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.810213] env[61243]: DEBUG nova.compute.manager [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Build of instance f21b7977-747e-437f-b03d-8442ed58756e was re-scheduled: Binding failed for port a5db5193-898e-47cb-8133-641566a114af, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 628.810213] env[61243]: DEBUG nova.compute.manager [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 628.810213] env[61243]: DEBUG oslo_concurrency.lockutils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquiring lock "refresh_cache-f21b7977-747e-437f-b03d-8442ed58756e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.810213] env[61243]: DEBUG oslo_concurrency.lockutils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquired lock "refresh_cache-f21b7977-747e-437f-b03d-8442ed58756e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.810416] env[61243]: DEBUG nova.network.neutron [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 628.843693] env[61243]: INFO nova.compute.manager [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] [instance: 2451843a-fd32-450e-933e-e635c789834b] Took 1.02 seconds to deallocate network for instance. [ 629.337809] env[61243]: DEBUG nova.network.neutron [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.483549] env[61243]: DEBUG nova.network.neutron [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.803241] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85143b84-afd2-49e8-afff-48b6f3876fd3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.811726] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f3d02e-0c01-49fe-934a-15488deb0a5e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.859649] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b630f8-9bdf-4ccc-bfae-e4b7fd7c0259 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.867793] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b383a29-b96b-45a5-9378-9e63fc283d2c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.884635] env[61243]: DEBUG nova.compute.provider_tree [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.886751] env[61243]: INFO nova.scheduler.client.report [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Deleted allocations for instance 2451843a-fd32-450e-933e-e635c789834b [ 629.989514] env[61243]: DEBUG oslo_concurrency.lockutils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Releasing lock "refresh_cache-f21b7977-747e-437f-b03d-8442ed58756e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.989764] env[61243]: DEBUG nova.compute.manager [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 629.989986] env[61243]: DEBUG nova.compute.manager [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.990861] env[61243]: DEBUG nova.network.neutron [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 630.007336] env[61243]: DEBUG nova.network.neutron [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.394923] env[61243]: DEBUG nova.scheduler.client.report [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 630.400246] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6669cb07-edf4-44b5-958a-93f2b4c2f4bf tempest-InstanceActionsV221TestJSON-1968650226 tempest-InstanceActionsV221TestJSON-1968650226-project-member] Lock "2451843a-fd32-450e-933e-e635c789834b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.350s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.511015] env[61243]: DEBUG nova.network.neutron [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.901644] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.096s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.902295] env[61243]: ERROR nova.compute.manager [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 432beef1-3f34-433c-9998-2c6210f10935, please check neutron logs for more information. [ 630.902295] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Traceback (most recent call last): [ 630.902295] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 630.902295] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] self.driver.spawn(context, instance, image_meta, [ 630.902295] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 630.902295] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.902295] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.902295] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] vm_ref = self.build_virtual_machine(instance, [ 630.902295] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.902295] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.902295] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.902597] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] for vif in network_info: [ 630.902597] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 630.902597] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] return self._sync_wrapper(fn, *args, **kwargs) [ 630.902597] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 630.902597] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] self.wait() [ 630.902597] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 630.902597] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] self[:] = self._gt.wait() [ 630.902597] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.902597] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] return self._exit_event.wait() [ 630.902597] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 630.902597] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] result = hub.switch() [ 630.902597] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 630.902597] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] return self.greenlet.switch() [ 630.902886] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.902886] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] result = function(*args, **kwargs) [ 630.902886] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 630.902886] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] return func(*args, **kwargs) [ 630.902886] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.902886] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] raise e [ 630.902886] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.902886] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] nwinfo = self.network_api.allocate_for_instance( [ 630.902886] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.902886] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] created_port_ids = self._update_ports_for_instance( [ 630.902886] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.902886] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] with excutils.save_and_reraise_exception(): [ 630.902886] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.903184] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] self.force_reraise() [ 630.903184] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.903184] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] raise self.value [ 630.903184] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.903184] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] updated_port = self._update_port( [ 630.903184] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.903184] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] _ensure_no_port_binding_failure(port) [ 630.903184] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.903184] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] raise exception.PortBindingFailed(port_id=port['id']) [ 630.903184] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] nova.exception.PortBindingFailed: Binding failed for port 432beef1-3f34-433c-9998-2c6210f10935, please check neutron logs for more information. [ 630.903184] env[61243]: ERROR nova.compute.manager [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] [ 630.903565] env[61243]: DEBUG nova.compute.utils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Binding failed for port 432beef1-3f34-433c-9998-2c6210f10935, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 630.904284] env[61243]: DEBUG nova.compute.manager [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 630.906842] env[61243]: DEBUG oslo_concurrency.lockutils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.898s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.909641] env[61243]: DEBUG nova.compute.manager [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Build of instance e2c498c5-3df1-4b99-a322-a598afd506f8 was re-scheduled: Binding failed for port 432beef1-3f34-433c-9998-2c6210f10935, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 630.910142] env[61243]: DEBUG nova.compute.manager [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 630.910379] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Acquiring lock "refresh_cache-e2c498c5-3df1-4b99-a322-a598afd506f8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.910530] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Acquired lock "refresh_cache-e2c498c5-3df1-4b99-a322-a598afd506f8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.910752] env[61243]: DEBUG nova.network.neutron [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.019636] env[61243]: INFO nova.compute.manager [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: f21b7977-747e-437f-b03d-8442ed58756e] Took 1.03 seconds to deallocate network for instance. [ 631.439364] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.453582] env[61243]: DEBUG nova.network.neutron [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.769350] env[61243]: DEBUG nova.network.neutron [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.905082] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e868d690-8d12-4a92-96dc-1c0f2fda1ca7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.913321] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17605d49-f6cd-405a-9fd1-32fd5dc58d9f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.948232] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babd9df5-a925-445c-81ed-3109a232fd06 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.955975] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b7684d4-3efa-4a8c-9cf6-5613c7cddcc1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.975902] env[61243]: DEBUG nova.compute.provider_tree [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.061643] env[61243]: INFO nova.scheduler.client.report [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Deleted allocations for instance f21b7977-747e-437f-b03d-8442ed58756e [ 632.273325] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Releasing lock "refresh_cache-e2c498c5-3df1-4b99-a322-a598afd506f8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.273602] env[61243]: DEBUG nova.compute.manager [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 632.273812] env[61243]: DEBUG nova.compute.manager [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.273986] env[61243]: DEBUG nova.network.neutron [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.303187] env[61243]: DEBUG nova.network.neutron [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.481991] env[61243]: DEBUG nova.scheduler.client.report [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.571048] env[61243]: DEBUG oslo_concurrency.lockutils [None req-afcf7feb-0a97-4492-8ea9-5b5b058b711d tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "f21b7977-747e-437f-b03d-8442ed58756e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.574s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.805588] env[61243]: DEBUG nova.network.neutron [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.987788] env[61243]: DEBUG oslo_concurrency.lockutils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.081s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.988598] env[61243]: ERROR nova.compute.manager [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7cc00845-7608-4748-9fca-d4628e324770, please check neutron logs for more information. [ 632.988598] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Traceback (most recent call last): [ 632.988598] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 632.988598] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] self.driver.spawn(context, instance, image_meta, [ 632.988598] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 632.988598] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.988598] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.988598] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] vm_ref = self.build_virtual_machine(instance, [ 632.988598] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.988598] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.988598] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.988923] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] for vif in network_info: [ 632.988923] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.988923] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] return self._sync_wrapper(fn, *args, **kwargs) [ 632.988923] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.988923] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] self.wait() [ 632.988923] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.988923] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] self[:] = self._gt.wait() [ 632.988923] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.988923] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] return self._exit_event.wait() [ 632.988923] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 632.988923] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] result = hub.switch() [ 632.988923] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 632.988923] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] return self.greenlet.switch() [ 632.989223] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.989223] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] result = function(*args, **kwargs) [ 632.989223] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 632.989223] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] return func(*args, **kwargs) [ 632.989223] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.989223] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] raise e [ 632.989223] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.989223] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] nwinfo = self.network_api.allocate_for_instance( [ 632.989223] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.989223] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] created_port_ids = self._update_ports_for_instance( [ 632.989223] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.989223] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] with excutils.save_and_reraise_exception(): [ 632.989223] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.989511] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] self.force_reraise() [ 632.989511] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.989511] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] raise self.value [ 632.989511] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.989511] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] updated_port = self._update_port( [ 632.989511] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.989511] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] _ensure_no_port_binding_failure(port) [ 632.989511] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.989511] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] raise exception.PortBindingFailed(port_id=port['id']) [ 632.989511] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] nova.exception.PortBindingFailed: Binding failed for port 7cc00845-7608-4748-9fca-d4628e324770, please check neutron logs for more information. [ 632.989511] env[61243]: ERROR nova.compute.manager [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] [ 632.989759] env[61243]: DEBUG nova.compute.utils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Binding failed for port 7cc00845-7608-4748-9fca-d4628e324770, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 632.991310] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.943s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.992829] env[61243]: INFO nova.compute.claims [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 632.996148] env[61243]: DEBUG nova.compute.manager [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Build of instance 5508a717-fb58-49b3-bfd1-b80d76a3c4e1 was re-scheduled: Binding failed for port 7cc00845-7608-4748-9fca-d4628e324770, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 632.996665] env[61243]: DEBUG nova.compute.manager [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 632.996920] env[61243]: DEBUG oslo_concurrency.lockutils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Acquiring lock "refresh_cache-5508a717-fb58-49b3-bfd1-b80d76a3c4e1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.997119] env[61243]: DEBUG oslo_concurrency.lockutils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Acquired lock "refresh_cache-5508a717-fb58-49b3-bfd1-b80d76a3c4e1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.997312] env[61243]: DEBUG nova.network.neutron [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 633.073792] env[61243]: DEBUG nova.compute.manager [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 633.310966] env[61243]: INFO nova.compute.manager [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] [instance: e2c498c5-3df1-4b99-a322-a598afd506f8] Took 1.03 seconds to deallocate network for instance. [ 633.524465] env[61243]: DEBUG nova.network.neutron [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.602406] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.686615] env[61243]: DEBUG nova.network.neutron [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.192806] env[61243]: DEBUG oslo_concurrency.lockutils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Releasing lock "refresh_cache-5508a717-fb58-49b3-bfd1-b80d76a3c4e1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.192875] env[61243]: DEBUG nova.compute.manager [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 634.192987] env[61243]: DEBUG nova.compute.manager [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.193598] env[61243]: DEBUG nova.network.neutron [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 634.217646] env[61243]: DEBUG nova.network.neutron [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.350173] env[61243]: INFO nova.scheduler.client.report [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Deleted allocations for instance e2c498c5-3df1-4b99-a322-a598afd506f8 [ 634.561886] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32b83f8-5afb-42d4-9851-b5003c2ce17e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.570299] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f011653-0be9-4e49-8f28-9dab749d3472 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.605441] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b66c45f6-4b09-41c2-b1f2-413d1a2986f5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.619926] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca38ec9-cc15-46dd-91e6-30e4140becc0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.634786] env[61243]: DEBUG nova.compute.provider_tree [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.721689] env[61243]: DEBUG nova.network.neutron [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.866531] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3853df46-e047-4000-8d30-9754b7f5600c tempest-FloatingIPsAssociationTestJSON-395431282 tempest-FloatingIPsAssociationTestJSON-395431282-project-member] Lock "e2c498c5-3df1-4b99-a322-a598afd506f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.804s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.140429] env[61243]: DEBUG nova.scheduler.client.report [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 635.226607] env[61243]: INFO nova.compute.manager [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 5508a717-fb58-49b3-bfd1-b80d76a3c4e1] Took 1.03 seconds to deallocate network for instance. [ 635.374386] env[61243]: DEBUG nova.compute.manager [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 635.646318] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.655s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.647893] env[61243]: DEBUG nova.compute.manager [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 635.651633] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.041s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.900627] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.156855] env[61243]: DEBUG nova.compute.utils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 636.162216] env[61243]: DEBUG nova.compute.manager [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 636.165314] env[61243]: DEBUG nova.network.neutron [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 636.233067] env[61243]: DEBUG nova.policy [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '61fcc31ce73a43cda9e8b304700c62be', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '867c187215a8441085a500e29ff49846', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 636.273454] env[61243]: INFO nova.scheduler.client.report [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Deleted allocations for instance 5508a717-fb58-49b3-bfd1-b80d76a3c4e1 [ 636.418059] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Acquiring lock "0d0504ae-8a82-4b5d-9d5d-535f30ceafa9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.419049] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Lock "0d0504ae-8a82-4b5d-9d5d-535f30ceafa9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.556394] env[61243]: DEBUG nova.network.neutron [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Successfully created port: 49584d24-a0d0-45b8-af7c-bf7f39fb1964 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 636.663155] env[61243]: DEBUG nova.compute.manager [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 636.696293] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c8377c5-c714-43eb-bf09-2de8f589b537 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.704877] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4bda1c1-bf12-4d93-914c-23ff233ae5d0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.739034] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea9c7eb-1dcc-4b08-9aab-2c48511788d1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.746905] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b9eb643-a8c0-4a54-97e2-cef6348c9900 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.761135] env[61243]: DEBUG nova.compute.provider_tree [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.790918] env[61243]: DEBUG oslo_concurrency.lockutils [None req-56dc689a-d7e1-451c-a749-d02cce8bae51 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Lock "5508a717-fb58-49b3-bfd1-b80d76a3c4e1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.725s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.266048] env[61243]: DEBUG nova.scheduler.client.report [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.294113] env[61243]: DEBUG nova.compute.manager [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 637.478249] env[61243]: DEBUG oslo_concurrency.lockutils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Acquiring lock "2a81aa94-3e30-4f0d-b540-1f98cfa4cea9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.478249] env[61243]: DEBUG oslo_concurrency.lockutils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Lock "2a81aa94-3e30-4f0d-b540-1f98cfa4cea9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.559027] env[61243]: DEBUG nova.compute.manager [req-02a77414-d723-4e09-9775-e7ca749c4dc3 req-b2024889-d791-4888-93d5-a72ea94d9fce service nova] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Received event network-changed-49584d24-a0d0-45b8-af7c-bf7f39fb1964 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 637.559229] env[61243]: DEBUG nova.compute.manager [req-02a77414-d723-4e09-9775-e7ca749c4dc3 req-b2024889-d791-4888-93d5-a72ea94d9fce service nova] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Refreshing instance network info cache due to event network-changed-49584d24-a0d0-45b8-af7c-bf7f39fb1964. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 637.559435] env[61243]: DEBUG oslo_concurrency.lockutils [req-02a77414-d723-4e09-9775-e7ca749c4dc3 req-b2024889-d791-4888-93d5-a72ea94d9fce service nova] Acquiring lock "refresh_cache-b3d98e6f-4b45-4eaf-a79c-b683b32fe053" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.559595] env[61243]: DEBUG oslo_concurrency.lockutils [req-02a77414-d723-4e09-9775-e7ca749c4dc3 req-b2024889-d791-4888-93d5-a72ea94d9fce service nova] Acquired lock "refresh_cache-b3d98e6f-4b45-4eaf-a79c-b683b32fe053" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.560056] env[61243]: DEBUG nova.network.neutron [req-02a77414-d723-4e09-9775-e7ca749c4dc3 req-b2024889-d791-4888-93d5-a72ea94d9fce service nova] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Refreshing network info cache for port 49584d24-a0d0-45b8-af7c-bf7f39fb1964 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 637.677543] env[61243]: DEBUG nova.compute.manager [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 637.712139] env[61243]: ERROR nova.compute.manager [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 49584d24-a0d0-45b8-af7c-bf7f39fb1964, please check neutron logs for more information. [ 637.712139] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 637.712139] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.712139] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 637.712139] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.712139] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 637.712139] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.712139] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 637.712139] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.712139] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 637.712139] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.712139] env[61243]: ERROR nova.compute.manager raise self.value [ 637.712139] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.712139] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 637.712139] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.712139] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 637.712659] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.712659] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 637.712659] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 49584d24-a0d0-45b8-af7c-bf7f39fb1964, please check neutron logs for more information. [ 637.712659] env[61243]: ERROR nova.compute.manager [ 637.712659] env[61243]: Traceback (most recent call last): [ 637.712659] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 637.712659] env[61243]: listener.cb(fileno) [ 637.712659] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.712659] env[61243]: result = function(*args, **kwargs) [ 637.712659] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 637.712659] env[61243]: return func(*args, **kwargs) [ 637.712659] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.712659] env[61243]: raise e [ 637.712659] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.712659] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 637.712659] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.712659] env[61243]: created_port_ids = self._update_ports_for_instance( [ 637.712659] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.712659] env[61243]: with excutils.save_and_reraise_exception(): [ 637.712659] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.712659] env[61243]: self.force_reraise() [ 637.712659] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.712659] env[61243]: raise self.value [ 637.712659] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.712659] env[61243]: updated_port = self._update_port( [ 637.712659] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.712659] env[61243]: _ensure_no_port_binding_failure(port) [ 637.712659] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.712659] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 637.713443] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 49584d24-a0d0-45b8-af7c-bf7f39fb1964, please check neutron logs for more information. [ 637.713443] env[61243]: Removing descriptor: 16 [ 637.716075] env[61243]: DEBUG nova.virt.hardware [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 637.716516] env[61243]: DEBUG nova.virt.hardware [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 637.717179] env[61243]: DEBUG nova.virt.hardware [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 637.717179] env[61243]: DEBUG nova.virt.hardware [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 637.717685] env[61243]: DEBUG nova.virt.hardware [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 637.718064] env[61243]: DEBUG nova.virt.hardware [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 637.720704] env[61243]: DEBUG nova.virt.hardware [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 637.720704] env[61243]: DEBUG nova.virt.hardware [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 637.720704] env[61243]: DEBUG nova.virt.hardware [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 637.720704] env[61243]: DEBUG nova.virt.hardware [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 637.720704] env[61243]: DEBUG nova.virt.hardware [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 637.720996] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cadf4b72-aed0-46f2-8940-1d9bef05fa7d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.734974] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed76a7fd-f6a5-450e-8c49-56f2332251e6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.751495] env[61243]: ERROR nova.compute.manager [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 49584d24-a0d0-45b8-af7c-bf7f39fb1964, please check neutron logs for more information. [ 637.751495] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Traceback (most recent call last): [ 637.751495] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 637.751495] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] yield resources [ 637.751495] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 637.751495] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] self.driver.spawn(context, instance, image_meta, [ 637.751495] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 637.751495] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.751495] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.751495] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] vm_ref = self.build_virtual_machine(instance, [ 637.751495] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.751864] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.751864] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.751864] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] for vif in network_info: [ 637.751864] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.751864] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] return self._sync_wrapper(fn, *args, **kwargs) [ 637.751864] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.751864] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] self.wait() [ 637.751864] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.751864] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] self[:] = self._gt.wait() [ 637.751864] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.751864] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] return self._exit_event.wait() [ 637.751864] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 637.751864] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] current.throw(*self._exc) [ 637.753018] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.753018] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] result = function(*args, **kwargs) [ 637.753018] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 637.753018] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] return func(*args, **kwargs) [ 637.753018] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.753018] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] raise e [ 637.753018] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.753018] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] nwinfo = self.network_api.allocate_for_instance( [ 637.753018] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.753018] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] created_port_ids = self._update_ports_for_instance( [ 637.753018] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.753018] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] with excutils.save_and_reraise_exception(): [ 637.753018] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.753392] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] self.force_reraise() [ 637.753392] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.753392] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] raise self.value [ 637.753392] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.753392] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] updated_port = self._update_port( [ 637.753392] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.753392] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] _ensure_no_port_binding_failure(port) [ 637.753392] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.753392] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] raise exception.PortBindingFailed(port_id=port['id']) [ 637.753392] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] nova.exception.PortBindingFailed: Binding failed for port 49584d24-a0d0-45b8-af7c-bf7f39fb1964, please check neutron logs for more information. [ 637.753392] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] [ 637.753392] env[61243]: INFO nova.compute.manager [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Terminating instance [ 637.754159] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Acquiring lock "refresh_cache-b3d98e6f-4b45-4eaf-a79c-b683b32fe053" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.774761] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.124s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.775478] env[61243]: ERROR nova.compute.manager [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 538d5952-f284-4e64-99ae-1378fa45eb8a, please check neutron logs for more information. [ 637.775478] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Traceback (most recent call last): [ 637.775478] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 637.775478] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] self.driver.spawn(context, instance, image_meta, [ 637.775478] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 637.775478] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.775478] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.775478] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] vm_ref = self.build_virtual_machine(instance, [ 637.775478] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.775478] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.775478] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.775832] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] for vif in network_info: [ 637.775832] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.775832] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] return self._sync_wrapper(fn, *args, **kwargs) [ 637.775832] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.775832] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] self.wait() [ 637.775832] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.775832] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] self[:] = self._gt.wait() [ 637.775832] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.775832] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] return self._exit_event.wait() [ 637.775832] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 637.775832] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] current.throw(*self._exc) [ 637.775832] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.775832] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] result = function(*args, **kwargs) [ 637.776150] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 637.776150] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] return func(*args, **kwargs) [ 637.776150] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.776150] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] raise e [ 637.776150] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.776150] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] nwinfo = self.network_api.allocate_for_instance( [ 637.776150] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.776150] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] created_port_ids = self._update_ports_for_instance( [ 637.776150] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.776150] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] with excutils.save_and_reraise_exception(): [ 637.776150] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.776150] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] self.force_reraise() [ 637.776150] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.776750] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] raise self.value [ 637.776750] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.776750] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] updated_port = self._update_port( [ 637.776750] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.776750] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] _ensure_no_port_binding_failure(port) [ 637.776750] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.776750] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] raise exception.PortBindingFailed(port_id=port['id']) [ 637.776750] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] nova.exception.PortBindingFailed: Binding failed for port 538d5952-f284-4e64-99ae-1378fa45eb8a, please check neutron logs for more information. [ 637.776750] env[61243]: ERROR nova.compute.manager [instance: 476632bb-e07c-4660-9ea4-7290884e4761] [ 637.776750] env[61243]: DEBUG nova.compute.utils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Binding failed for port 538d5952-f284-4e64-99ae-1378fa45eb8a, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 637.777287] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.833s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.779798] env[61243]: INFO nova.compute.claims [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 637.781830] env[61243]: DEBUG nova.compute.manager [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Build of instance 476632bb-e07c-4660-9ea4-7290884e4761 was re-scheduled: Binding failed for port 538d5952-f284-4e64-99ae-1378fa45eb8a, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 637.782297] env[61243]: DEBUG nova.compute.manager [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 637.782721] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquiring lock "refresh_cache-476632bb-e07c-4660-9ea4-7290884e4761" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.782721] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquired lock "refresh_cache-476632bb-e07c-4660-9ea4-7290884e4761" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.783996] env[61243]: DEBUG nova.network.neutron [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 637.822841] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.046894] env[61243]: DEBUG nova.compute.manager [None req-888bce62-a6c3-4f99-8564-7116a72a7895 tempest-ServerDiagnosticsV248Test-942551326 tempest-ServerDiagnosticsV248Test-942551326-project-admin] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 638.049582] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfdfa909-a22c-4cfd-9065-cc0e62418492 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.060484] env[61243]: INFO nova.compute.manager [None req-888bce62-a6c3-4f99-8564-7116a72a7895 tempest-ServerDiagnosticsV248Test-942551326 tempest-ServerDiagnosticsV248Test-942551326-project-admin] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Retrieving diagnostics [ 638.061541] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89965142-db39-461c-b599-5900d1be371a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.097105] env[61243]: DEBUG nova.network.neutron [req-02a77414-d723-4e09-9775-e7ca749c4dc3 req-b2024889-d791-4888-93d5-a72ea94d9fce service nova] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.231516] env[61243]: DEBUG nova.network.neutron [req-02a77414-d723-4e09-9775-e7ca749c4dc3 req-b2024889-d791-4888-93d5-a72ea94d9fce service nova] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.306676] env[61243]: DEBUG nova.network.neutron [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.429125] env[61243]: DEBUG nova.network.neutron [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.733652] env[61243]: DEBUG oslo_concurrency.lockutils [req-02a77414-d723-4e09-9775-e7ca749c4dc3 req-b2024889-d791-4888-93d5-a72ea94d9fce service nova] Releasing lock "refresh_cache-b3d98e6f-4b45-4eaf-a79c-b683b32fe053" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.734154] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Acquired lock "refresh_cache-b3d98e6f-4b45-4eaf-a79c-b683b32fe053" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.734258] env[61243]: DEBUG nova.network.neutron [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.932044] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Releasing lock "refresh_cache-476632bb-e07c-4660-9ea4-7290884e4761" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.932294] env[61243]: DEBUG nova.compute.manager [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 638.932478] env[61243]: DEBUG nova.compute.manager [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 638.932643] env[61243]: DEBUG nova.network.neutron [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 638.954110] env[61243]: DEBUG nova.network.neutron [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.215357] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Acquiring lock "b5163d89-ceb8-4c61-ae02-0ae3311b58a9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.215695] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Lock "b5163d89-ceb8-4c61-ae02-0ae3311b58a9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.215862] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Acquiring lock "b5163d89-ceb8-4c61-ae02-0ae3311b58a9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.216053] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Lock "b5163d89-ceb8-4c61-ae02-0ae3311b58a9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.216238] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Lock "b5163d89-ceb8-4c61-ae02-0ae3311b58a9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.219594] env[61243]: INFO nova.compute.manager [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Terminating instance [ 639.221612] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Acquiring lock "refresh_cache-b5163d89-ceb8-4c61-ae02-0ae3311b58a9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.221612] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Acquired lock "refresh_cache-b5163d89-ceb8-4c61-ae02-0ae3311b58a9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.221770] env[61243]: DEBUG nova.network.neutron [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 639.231015] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f199a21e-8ba1-47c9-928a-424c2dd5079e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.242914] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68fa8be4-af40-4412-b054-04c52d6246e7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.275178] env[61243]: DEBUG nova.network.neutron [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.277266] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43cbb65-916f-41c7-a08c-91897a9a1d14 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.285926] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c933aa2-b5ab-491d-a6b8-08634a769fb9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.299996] env[61243]: DEBUG nova.compute.provider_tree [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.355113] env[61243]: DEBUG nova.network.neutron [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.457079] env[61243]: DEBUG nova.network.neutron [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.635749] env[61243]: DEBUG nova.compute.manager [req-fc166228-017e-4032-8957-19e75f6aa0cc req-a62ccb95-82a0-4136-b6a8-72250240a113 service nova] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Received event network-vif-deleted-49584d24-a0d0-45b8-af7c-bf7f39fb1964 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 639.741181] env[61243]: DEBUG nova.network.neutron [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.795639] env[61243]: DEBUG nova.network.neutron [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.806591] env[61243]: DEBUG nova.scheduler.client.report [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.862896] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Releasing lock "refresh_cache-b3d98e6f-4b45-4eaf-a79c-b683b32fe053" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.862896] env[61243]: DEBUG nova.compute.manager [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 639.862896] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 639.862896] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a808661c-75e7-4e35-ae9e-101ff2c38444 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.872813] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8539da3-5319-4e7a-b72d-b90fe28a9412 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.898884] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b3d98e6f-4b45-4eaf-a79c-b683b32fe053 could not be found. [ 639.899246] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 639.899649] env[61243]: INFO nova.compute.manager [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Took 0.04 seconds to destroy the instance on the hypervisor. [ 639.899942] env[61243]: DEBUG oslo.service.loopingcall [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 639.900254] env[61243]: DEBUG nova.compute.manager [-] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.900405] env[61243]: DEBUG nova.network.neutron [-] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.917436] env[61243]: DEBUG nova.network.neutron [-] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.960281] env[61243]: INFO nova.compute.manager [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 476632bb-e07c-4660-9ea4-7290884e4761] Took 1.03 seconds to deallocate network for instance. [ 640.267448] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Acquiring lock "1f4985a7-f6cc-4e79-aea3-ce0623aed515" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.267688] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Lock "1f4985a7-f6cc-4e79-aea3-ce0623aed515" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.297549] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Releasing lock "refresh_cache-b5163d89-ceb8-4c61-ae02-0ae3311b58a9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.298774] env[61243]: DEBUG nova.compute.manager [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 640.298774] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 640.299159] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac7b892-71c1-43da-8959-4e8ae63ca20e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.307664] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 640.307922] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-63e20ec4-f93f-4eb0-9598-fc622ce439ba {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.312305] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.535s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.312811] env[61243]: DEBUG nova.compute.manager [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 640.316877] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.823s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.319690] env[61243]: DEBUG oslo_vmware.api [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Waiting for the task: (returnval){ [ 640.319690] env[61243]: value = "task-1338744" [ 640.319690] env[61243]: _type = "Task" [ 640.319690] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.330046] env[61243]: DEBUG oslo_vmware.api [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': task-1338744, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.419943] env[61243]: DEBUG nova.network.neutron [-] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.821378] env[61243]: DEBUG nova.compute.utils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 640.826863] env[61243]: DEBUG nova.compute.manager [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 640.827060] env[61243]: DEBUG nova.network.neutron [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 640.840147] env[61243]: DEBUG oslo_vmware.api [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': task-1338744, 'name': PowerOffVM_Task, 'duration_secs': 0.120755} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.840329] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 640.840541] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 640.840739] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e65182dd-3ae0-4512-ad00-fbb51a26bb28 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.870556] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 640.870556] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 640.870556] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Deleting the datastore file [datastore2] b5163d89-ceb8-4c61-ae02-0ae3311b58a9 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 640.870556] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b92b692a-a148-44b1-a2cf-7ab935189fa5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.875589] env[61243]: DEBUG oslo_vmware.api [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Waiting for the task: (returnval){ [ 640.875589] env[61243]: value = "task-1338746" [ 640.875589] env[61243]: _type = "Task" [ 640.875589] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.879689] env[61243]: DEBUG nova.policy [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bb353472a60941228b74f8b35546a2df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c116557d493340aebe37e22d1e7fe003', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 640.889582] env[61243]: DEBUG oslo_vmware.api [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': task-1338746, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.926751] env[61243]: INFO nova.compute.manager [-] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Took 1.03 seconds to deallocate network for instance. [ 640.929417] env[61243]: DEBUG nova.compute.claims [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 640.929604] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.990087] env[61243]: INFO nova.scheduler.client.report [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Deleted allocations for instance 476632bb-e07c-4660-9ea4-7290884e4761 [ 641.164370] env[61243]: DEBUG nova.network.neutron [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Successfully created port: f3056083-ab3a-48ad-ae6c-37e5fa1b3d60 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 641.233470] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14edcff7-700f-477a-bad4-b4e7c02bbf4b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.241484] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c25f6a1-c544-487c-8d52-ff93fb91c784 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.270233] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a93d41a5-bd60-4714-84a7-ed996553a489 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.278685] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03bfb241-769d-4cc0-a362-925c70f1f30c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.293204] env[61243]: DEBUG nova.compute.provider_tree [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.327488] env[61243]: DEBUG nova.compute.manager [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 641.387406] env[61243]: DEBUG oslo_vmware.api [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Task: {'id': task-1338746, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095312} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.387949] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 641.388210] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 641.388624] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 641.388874] env[61243]: INFO nova.compute.manager [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Took 1.09 seconds to destroy the instance on the hypervisor. [ 641.389163] env[61243]: DEBUG oslo.service.loopingcall [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 641.389412] env[61243]: DEBUG nova.compute.manager [-] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 641.389719] env[61243]: DEBUG nova.network.neutron [-] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 641.407365] env[61243]: DEBUG nova.network.neutron [-] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.500478] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41c1bc4c-8d4b-40af-ab12-986a2d89606a tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "476632bb-e07c-4660-9ea4-7290884e4761" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.756s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.604951] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 641.605192] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 641.796445] env[61243]: DEBUG nova.scheduler.client.report [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.911796] env[61243]: DEBUG nova.network.neutron [-] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.958345] env[61243]: DEBUG nova.compute.manager [req-5a8b552d-e748-45aa-8763-cea19e3a4364 req-e4d1dd4a-ba7c-49c7-a3ef-dfe7d0153107 service nova] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Received event network-changed-f3056083-ab3a-48ad-ae6c-37e5fa1b3d60 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 641.958763] env[61243]: DEBUG nova.compute.manager [req-5a8b552d-e748-45aa-8763-cea19e3a4364 req-e4d1dd4a-ba7c-49c7-a3ef-dfe7d0153107 service nova] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Refreshing instance network info cache due to event network-changed-f3056083-ab3a-48ad-ae6c-37e5fa1b3d60. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 641.958763] env[61243]: DEBUG oslo_concurrency.lockutils [req-5a8b552d-e748-45aa-8763-cea19e3a4364 req-e4d1dd4a-ba7c-49c7-a3ef-dfe7d0153107 service nova] Acquiring lock "refresh_cache-0b223791-7e95-43fe-bef0-bac6222884ea" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.958893] env[61243]: DEBUG oslo_concurrency.lockutils [req-5a8b552d-e748-45aa-8763-cea19e3a4364 req-e4d1dd4a-ba7c-49c7-a3ef-dfe7d0153107 service nova] Acquired lock "refresh_cache-0b223791-7e95-43fe-bef0-bac6222884ea" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.959103] env[61243]: DEBUG nova.network.neutron [req-5a8b552d-e748-45aa-8763-cea19e3a4364 req-e4d1dd4a-ba7c-49c7-a3ef-dfe7d0153107 service nova] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Refreshing network info cache for port f3056083-ab3a-48ad-ae6c-37e5fa1b3d60 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 642.003631] env[61243]: DEBUG nova.compute.manager [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 642.111969] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 642.111969] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Starting heal instance info cache {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 642.111969] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Rebuilding the list of instances to heal {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 642.140918] env[61243]: ERROR nova.compute.manager [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f3056083-ab3a-48ad-ae6c-37e5fa1b3d60, please check neutron logs for more information. [ 642.140918] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 642.140918] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.140918] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 642.140918] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.140918] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 642.140918] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.140918] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 642.140918] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.140918] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 642.140918] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.140918] env[61243]: ERROR nova.compute.manager raise self.value [ 642.140918] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.140918] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 642.140918] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.140918] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 642.141308] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.141308] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 642.141308] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f3056083-ab3a-48ad-ae6c-37e5fa1b3d60, please check neutron logs for more information. [ 642.141308] env[61243]: ERROR nova.compute.manager [ 642.141308] env[61243]: Traceback (most recent call last): [ 642.141308] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 642.141308] env[61243]: listener.cb(fileno) [ 642.141308] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.141308] env[61243]: result = function(*args, **kwargs) [ 642.141308] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 642.141308] env[61243]: return func(*args, **kwargs) [ 642.141308] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.141308] env[61243]: raise e [ 642.141308] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.141308] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 642.141308] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.141308] env[61243]: created_port_ids = self._update_ports_for_instance( [ 642.141308] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.141308] env[61243]: with excutils.save_and_reraise_exception(): [ 642.141308] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.141308] env[61243]: self.force_reraise() [ 642.141308] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.141308] env[61243]: raise self.value [ 642.141308] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.141308] env[61243]: updated_port = self._update_port( [ 642.141308] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.141308] env[61243]: _ensure_no_port_binding_failure(port) [ 642.141308] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.141308] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 642.142033] env[61243]: nova.exception.PortBindingFailed: Binding failed for port f3056083-ab3a-48ad-ae6c-37e5fa1b3d60, please check neutron logs for more information. [ 642.142033] env[61243]: Removing descriptor: 16 [ 642.302116] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.985s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.302762] env[61243]: ERROR nova.compute.manager [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 740571e3-5361-4bd1-abce-b7699711237f, please check neutron logs for more information. [ 642.302762] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Traceback (most recent call last): [ 642.302762] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.302762] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] self.driver.spawn(context, instance, image_meta, [ 642.302762] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 642.302762] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.302762] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.302762] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] vm_ref = self.build_virtual_machine(instance, [ 642.302762] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.302762] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.302762] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.303122] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] for vif in network_info: [ 642.303122] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.303122] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] return self._sync_wrapper(fn, *args, **kwargs) [ 642.303122] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.303122] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] self.wait() [ 642.303122] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.303122] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] self[:] = self._gt.wait() [ 642.303122] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.303122] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] return self._exit_event.wait() [ 642.303122] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 642.303122] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] current.throw(*self._exc) [ 642.303122] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.303122] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] result = function(*args, **kwargs) [ 642.303464] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 642.303464] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] return func(*args, **kwargs) [ 642.303464] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.303464] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] raise e [ 642.303464] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.303464] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] nwinfo = self.network_api.allocate_for_instance( [ 642.303464] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.303464] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] created_port_ids = self._update_ports_for_instance( [ 642.303464] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.303464] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] with excutils.save_and_reraise_exception(): [ 642.303464] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.303464] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] self.force_reraise() [ 642.303464] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.303821] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] raise self.value [ 642.303821] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.303821] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] updated_port = self._update_port( [ 642.303821] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.303821] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] _ensure_no_port_binding_failure(port) [ 642.303821] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.303821] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] raise exception.PortBindingFailed(port_id=port['id']) [ 642.303821] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] nova.exception.PortBindingFailed: Binding failed for port 740571e3-5361-4bd1-abce-b7699711237f, please check neutron logs for more information. [ 642.303821] env[61243]: ERROR nova.compute.manager [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] [ 642.303821] env[61243]: DEBUG nova.compute.utils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Binding failed for port 740571e3-5361-4bd1-abce-b7699711237f, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 642.304719] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.985s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.306267] env[61243]: INFO nova.compute.claims [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 642.308875] env[61243]: DEBUG nova.compute.manager [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Build of instance 461b2ef2-d665-47ab-affa-e525bffe561a was re-scheduled: Binding failed for port 740571e3-5361-4bd1-abce-b7699711237f, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 642.309289] env[61243]: DEBUG nova.compute.manager [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 642.309515] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquiring lock "refresh_cache-461b2ef2-d665-47ab-affa-e525bffe561a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.309662] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Acquired lock "refresh_cache-461b2ef2-d665-47ab-affa-e525bffe561a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.309820] env[61243]: DEBUG nova.network.neutron [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 642.337266] env[61243]: DEBUG nova.compute.manager [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 642.361714] env[61243]: DEBUG nova.virt.hardware [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 642.361714] env[61243]: DEBUG nova.virt.hardware [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 642.361938] env[61243]: DEBUG nova.virt.hardware [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 642.362035] env[61243]: DEBUG nova.virt.hardware [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 642.362191] env[61243]: DEBUG nova.virt.hardware [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 642.362364] env[61243]: DEBUG nova.virt.hardware [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 642.362582] env[61243]: DEBUG nova.virt.hardware [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 642.362739] env[61243]: DEBUG nova.virt.hardware [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 642.362907] env[61243]: DEBUG nova.virt.hardware [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 642.363175] env[61243]: DEBUG nova.virt.hardware [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 642.363375] env[61243]: DEBUG nova.virt.hardware [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 642.364225] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ac519e-77d0-494a-a778-adb614cc313d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.372206] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca0e6f1-be12-41f2-b63f-24ef18e9488b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.385502] env[61243]: ERROR nova.compute.manager [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f3056083-ab3a-48ad-ae6c-37e5fa1b3d60, please check neutron logs for more information. [ 642.385502] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Traceback (most recent call last): [ 642.385502] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 642.385502] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] yield resources [ 642.385502] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.385502] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] self.driver.spawn(context, instance, image_meta, [ 642.385502] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 642.385502] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.385502] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.385502] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] vm_ref = self.build_virtual_machine(instance, [ 642.385502] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.385841] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.385841] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.385841] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] for vif in network_info: [ 642.385841] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.385841] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] return self._sync_wrapper(fn, *args, **kwargs) [ 642.385841] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.385841] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] self.wait() [ 642.385841] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.385841] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] self[:] = self._gt.wait() [ 642.385841] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.385841] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] return self._exit_event.wait() [ 642.385841] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 642.385841] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] current.throw(*self._exc) [ 642.386154] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.386154] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] result = function(*args, **kwargs) [ 642.386154] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 642.386154] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] return func(*args, **kwargs) [ 642.386154] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.386154] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] raise e [ 642.386154] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.386154] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] nwinfo = self.network_api.allocate_for_instance( [ 642.386154] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.386154] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] created_port_ids = self._update_ports_for_instance( [ 642.386154] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.386154] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] with excutils.save_and_reraise_exception(): [ 642.386154] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.386519] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] self.force_reraise() [ 642.386519] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.386519] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] raise self.value [ 642.386519] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.386519] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] updated_port = self._update_port( [ 642.386519] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.386519] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] _ensure_no_port_binding_failure(port) [ 642.386519] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.386519] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] raise exception.PortBindingFailed(port_id=port['id']) [ 642.386519] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] nova.exception.PortBindingFailed: Binding failed for port f3056083-ab3a-48ad-ae6c-37e5fa1b3d60, please check neutron logs for more information. [ 642.386519] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] [ 642.386519] env[61243]: INFO nova.compute.manager [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Terminating instance [ 642.387576] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Acquiring lock "refresh_cache-0b223791-7e95-43fe-bef0-bac6222884ea" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.414480] env[61243]: INFO nova.compute.manager [-] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Took 1.02 seconds to deallocate network for instance. [ 642.476190] env[61243]: DEBUG nova.network.neutron [req-5a8b552d-e748-45aa-8763-cea19e3a4364 req-e4d1dd4a-ba7c-49c7-a3ef-dfe7d0153107 service nova] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.526757] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.548868] env[61243]: DEBUG nova.network.neutron [req-5a8b552d-e748-45aa-8763-cea19e3a4364 req-e4d1dd4a-ba7c-49c7-a3ef-dfe7d0153107 service nova] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.614620] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Skipping network cache update for instance because it is being deleted. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 642.614801] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Skipping network cache update for instance because it is Building. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 642.614933] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Skipping network cache update for instance because it is Building. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 642.615074] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Didn't find any instances for network info cache update. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 642.615269] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 642.615447] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 642.615570] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 642.615717] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 642.615871] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 642.616031] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 642.616166] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61243) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 642.616316] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager.update_available_resource {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 642.831098] env[61243]: DEBUG nova.network.neutron [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.922995] env[61243]: DEBUG nova.network.neutron [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.922995] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.053805] env[61243]: DEBUG oslo_concurrency.lockutils [req-5a8b552d-e748-45aa-8763-cea19e3a4364 req-e4d1dd4a-ba7c-49c7-a3ef-dfe7d0153107 service nova] Releasing lock "refresh_cache-0b223791-7e95-43fe-bef0-bac6222884ea" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.053805] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Acquired lock "refresh_cache-0b223791-7e95-43fe-bef0-bac6222884ea" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.053805] env[61243]: DEBUG nova.network.neutron [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 643.119075] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.424139] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Releasing lock "refresh_cache-461b2ef2-d665-47ab-affa-e525bffe561a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.424376] env[61243]: DEBUG nova.compute.manager [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 643.424565] env[61243]: DEBUG nova.compute.manager [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.424724] env[61243]: DEBUG nova.network.neutron [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.441697] env[61243]: DEBUG nova.network.neutron [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.570608] env[61243]: DEBUG nova.network.neutron [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.633475] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19c6794-e952-4eaf-b123-85c81bee61b4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.641215] env[61243]: DEBUG nova.network.neutron [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.643013] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0653980e-c092-44ac-946c-00ca64a4bc61 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.672700] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e0cbf1d-888e-4e9a-94f6-cf6645c76631 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.680374] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab176bb-0a1a-4437-9f1b-431644643b61 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.693607] env[61243]: DEBUG nova.compute.provider_tree [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.944718] env[61243]: DEBUG nova.network.neutron [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.009778] env[61243]: DEBUG nova.compute.manager [req-5c01ff25-96a0-4c1c-9f41-591384dbdf0d req-5cd6fa38-632b-46bb-91f3-3cc13c451916 service nova] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Received event network-vif-deleted-f3056083-ab3a-48ad-ae6c-37e5fa1b3d60 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 644.147199] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Releasing lock "refresh_cache-0b223791-7e95-43fe-bef0-bac6222884ea" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.147640] env[61243]: DEBUG nova.compute.manager [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 644.147837] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 644.148180] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7f60a2c8-ea99-44db-98cd-b355e86de52e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.157938] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f847d2f5-d365-48c1-b129-8bb2d09cb88a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.178196] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0b223791-7e95-43fe-bef0-bac6222884ea could not be found. [ 644.178436] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 644.178616] env[61243]: INFO nova.compute.manager [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Took 0.03 seconds to destroy the instance on the hypervisor. [ 644.179026] env[61243]: DEBUG oslo.service.loopingcall [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 644.179133] env[61243]: DEBUG nova.compute.manager [-] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 644.179175] env[61243]: DEBUG nova.network.neutron [-] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 644.193925] env[61243]: DEBUG nova.network.neutron [-] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.197059] env[61243]: DEBUG nova.scheduler.client.report [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 644.449169] env[61243]: INFO nova.compute.manager [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] [instance: 461b2ef2-d665-47ab-affa-e525bffe561a] Took 1.02 seconds to deallocate network for instance. [ 644.696365] env[61243]: DEBUG nova.network.neutron [-] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.701877] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.702266] env[61243]: DEBUG nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 644.705294] env[61243]: DEBUG oslo_concurrency.lockutils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.900s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.706695] env[61243]: INFO nova.compute.claims [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.200058] env[61243]: INFO nova.compute.manager [-] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Took 1.02 seconds to deallocate network for instance. [ 645.202084] env[61243]: DEBUG nova.compute.claims [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 645.202255] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.210660] env[61243]: DEBUG nova.compute.utils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 645.214064] env[61243]: DEBUG nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 645.214161] env[61243]: DEBUG nova.network.neutron [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 645.255322] env[61243]: DEBUG nova.policy [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d62b775c67c4fe5be1ed29156a164f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fd5607c942294bac94a3a6fd0fc981c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 645.474867] env[61243]: INFO nova.scheduler.client.report [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Deleted allocations for instance 461b2ef2-d665-47ab-affa-e525bffe561a [ 645.525040] env[61243]: DEBUG nova.network.neutron [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Successfully created port: 66e28bb3-150d-4a9f-8b17-b7c1911a40cd {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 645.721288] env[61243]: DEBUG nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 645.986205] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4b666c95-6a4d-482a-b1dc-96bf5c9f4118 tempest-ListServerFiltersTestJSON-312585624 tempest-ListServerFiltersTestJSON-312585624-project-member] Lock "461b2ef2-d665-47ab-affa-e525bffe561a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.230s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.111057] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cd37c8f-17c3-4895-8cb4-11bd12c00d0a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.118893] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c25059e-11fe-44e7-b7bf-f918f8d0cc63 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.148252] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b510707f-2640-4f8e-8946-8f217edcf897 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.158052] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f073fdf5-75c1-43fe-88e9-17c735f6339c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.168612] env[61243]: DEBUG nova.compute.provider_tree [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.200134] env[61243]: DEBUG nova.compute.manager [req-b009d706-3070-4447-93c1-9543efde48cb req-e35c7ae5-1ab7-46b6-9c8b-d197ef649072 service nova] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Received event network-changed-66e28bb3-150d-4a9f-8b17-b7c1911a40cd {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 646.200466] env[61243]: DEBUG nova.compute.manager [req-b009d706-3070-4447-93c1-9543efde48cb req-e35c7ae5-1ab7-46b6-9c8b-d197ef649072 service nova] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Refreshing instance network info cache due to event network-changed-66e28bb3-150d-4a9f-8b17-b7c1911a40cd. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 646.200590] env[61243]: DEBUG oslo_concurrency.lockutils [req-b009d706-3070-4447-93c1-9543efde48cb req-e35c7ae5-1ab7-46b6-9c8b-d197ef649072 service nova] Acquiring lock "refresh_cache-0625ac10-55b4-488a-aba5-bdaa394045f4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.200724] env[61243]: DEBUG oslo_concurrency.lockutils [req-b009d706-3070-4447-93c1-9543efde48cb req-e35c7ae5-1ab7-46b6-9c8b-d197ef649072 service nova] Acquired lock "refresh_cache-0625ac10-55b4-488a-aba5-bdaa394045f4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.200881] env[61243]: DEBUG nova.network.neutron [req-b009d706-3070-4447-93c1-9543efde48cb req-e35c7ae5-1ab7-46b6-9c8b-d197ef649072 service nova] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Refreshing network info cache for port 66e28bb3-150d-4a9f-8b17-b7c1911a40cd {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 646.226286] env[61243]: INFO nova.virt.block_device [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Booting with volume eff68297-bd34-4507-8de3-2dd6fe9d4265 at /dev/sda [ 646.274127] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-095f9471-f3dd-460d-aea2-c1044dc0033b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.283625] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddadc8a0-0005-4c04-99e7-c9f88754cde7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.305141] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-48da5db9-37f2-466a-9e67-be2c6495c016 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.312586] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20bf46c7-f3f6-4d63-9895-69002c6e4a60 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.336630] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c207e898-9b91-43f7-8af8-b9a22a7d25c0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.343879] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ff3e93-a4ca-4e71-a91b-7a4f9330cda2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.356828] env[61243]: DEBUG nova.virt.block_device [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Updating existing volume attachment record: 51526c2e-2c3f-44fe-b370-a00568491868 {{(pid=61243) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 646.382930] env[61243]: ERROR nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 66e28bb3-150d-4a9f-8b17-b7c1911a40cd, please check neutron logs for more information. [ 646.382930] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 646.382930] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.382930] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 646.382930] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.382930] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 646.382930] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.382930] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 646.382930] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.382930] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 646.382930] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.382930] env[61243]: ERROR nova.compute.manager raise self.value [ 646.382930] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.382930] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 646.382930] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.382930] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 646.383415] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.383415] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 646.383415] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 66e28bb3-150d-4a9f-8b17-b7c1911a40cd, please check neutron logs for more information. [ 646.383415] env[61243]: ERROR nova.compute.manager [ 646.383415] env[61243]: Traceback (most recent call last): [ 646.383415] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 646.383415] env[61243]: listener.cb(fileno) [ 646.383415] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.383415] env[61243]: result = function(*args, **kwargs) [ 646.383415] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 646.383415] env[61243]: return func(*args, **kwargs) [ 646.383415] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.383415] env[61243]: raise e [ 646.383415] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.383415] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 646.383415] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.383415] env[61243]: created_port_ids = self._update_ports_for_instance( [ 646.383415] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.383415] env[61243]: with excutils.save_and_reraise_exception(): [ 646.383415] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.383415] env[61243]: self.force_reraise() [ 646.383415] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.383415] env[61243]: raise self.value [ 646.383415] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.383415] env[61243]: updated_port = self._update_port( [ 646.383415] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.383415] env[61243]: _ensure_no_port_binding_failure(port) [ 646.383415] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.383415] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 646.384297] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 66e28bb3-150d-4a9f-8b17-b7c1911a40cd, please check neutron logs for more information. [ 646.384297] env[61243]: Removing descriptor: 16 [ 646.489674] env[61243]: DEBUG nova.compute.manager [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 646.673062] env[61243]: DEBUG nova.scheduler.client.report [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 646.720096] env[61243]: DEBUG nova.network.neutron [req-b009d706-3070-4447-93c1-9543efde48cb req-e35c7ae5-1ab7-46b6-9c8b-d197ef649072 service nova] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.793619] env[61243]: DEBUG nova.network.neutron [req-b009d706-3070-4447-93c1-9543efde48cb req-e35c7ae5-1ab7-46b6-9c8b-d197ef649072 service nova] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.010922] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.178902] env[61243]: DEBUG oslo_concurrency.lockutils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.179558] env[61243]: DEBUG nova.compute.manager [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 647.182733] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.744s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.184245] env[61243]: INFO nova.compute.claims [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 647.298198] env[61243]: DEBUG oslo_concurrency.lockutils [req-b009d706-3070-4447-93c1-9543efde48cb req-e35c7ae5-1ab7-46b6-9c8b-d197ef649072 service nova] Releasing lock "refresh_cache-0625ac10-55b4-488a-aba5-bdaa394045f4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.695023] env[61243]: DEBUG nova.compute.utils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 647.695262] env[61243]: DEBUG nova.compute.manager [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 647.695396] env[61243]: DEBUG nova.network.neutron [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 647.745833] env[61243]: DEBUG nova.policy [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '651117965b7b42da875b449e4d0b81af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '20ae73e7de094bafa1f423d9e90f770c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 648.033010] env[61243]: DEBUG nova.network.neutron [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Successfully created port: e148ba07-8dcb-4345-93fa-15a43fbbbe3f {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 648.198835] env[61243]: DEBUG nova.compute.manager [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 648.238628] env[61243]: DEBUG nova.compute.manager [req-46224dc1-d437-485b-b38e-4a455f685e8d req-7edbea6f-6011-456b-96e3-f271b6784f07 service nova] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Received event network-vif-deleted-66e28bb3-150d-4a9f-8b17-b7c1911a40cd {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 648.468645] env[61243]: DEBUG nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 648.469157] env[61243]: DEBUG nova.virt.hardware [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 648.469364] env[61243]: DEBUG nova.virt.hardware [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 648.469513] env[61243]: DEBUG nova.virt.hardware [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 648.469686] env[61243]: DEBUG nova.virt.hardware [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 648.469826] env[61243]: DEBUG nova.virt.hardware [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 648.469967] env[61243]: DEBUG nova.virt.hardware [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 648.470178] env[61243]: DEBUG nova.virt.hardware [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 648.470365] env[61243]: DEBUG nova.virt.hardware [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 648.470530] env[61243]: DEBUG nova.virt.hardware [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 648.470688] env[61243]: DEBUG nova.virt.hardware [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 648.470856] env[61243]: DEBUG nova.virt.hardware [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 648.471728] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97cc7012-9afa-4ab6-a27e-7cf72f95a855 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.481525] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89b2bfe9-b258-45c9-a6f8-e6b280af475c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.496998] env[61243]: ERROR nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 66e28bb3-150d-4a9f-8b17-b7c1911a40cd, please check neutron logs for more information. [ 648.496998] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Traceback (most recent call last): [ 648.496998] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 648.496998] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] yield resources [ 648.496998] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.496998] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] self.driver.spawn(context, instance, image_meta, [ 648.496998] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 648.496998] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.496998] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.496998] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] vm_ref = self.build_virtual_machine(instance, [ 648.496998] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.497377] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.497377] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.497377] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] for vif in network_info: [ 648.497377] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.497377] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] return self._sync_wrapper(fn, *args, **kwargs) [ 648.497377] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.497377] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] self.wait() [ 648.497377] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.497377] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] self[:] = self._gt.wait() [ 648.497377] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.497377] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] return self._exit_event.wait() [ 648.497377] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 648.497377] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] current.throw(*self._exc) [ 648.497741] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.497741] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] result = function(*args, **kwargs) [ 648.497741] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 648.497741] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] return func(*args, **kwargs) [ 648.497741] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.497741] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] raise e [ 648.497741] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.497741] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] nwinfo = self.network_api.allocate_for_instance( [ 648.497741] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.497741] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] created_port_ids = self._update_ports_for_instance( [ 648.497741] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.497741] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] with excutils.save_and_reraise_exception(): [ 648.497741] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.498136] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] self.force_reraise() [ 648.498136] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.498136] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] raise self.value [ 648.498136] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.498136] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] updated_port = self._update_port( [ 648.498136] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.498136] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] _ensure_no_port_binding_failure(port) [ 648.498136] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.498136] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] raise exception.PortBindingFailed(port_id=port['id']) [ 648.498136] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] nova.exception.PortBindingFailed: Binding failed for port 66e28bb3-150d-4a9f-8b17-b7c1911a40cd, please check neutron logs for more information. [ 648.498136] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] [ 648.498136] env[61243]: INFO nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Terminating instance [ 648.504438] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Acquiring lock "refresh_cache-0625ac10-55b4-488a-aba5-bdaa394045f4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.504438] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Acquired lock "refresh_cache-0625ac10-55b4-488a-aba5-bdaa394045f4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.504438] env[61243]: DEBUG nova.network.neutron [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 648.571835] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef062e9-d093-4915-b75a-5b2a9710305d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.580436] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8974bf28-82e9-428d-96a2-e2d014ffdd99 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.612662] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74fb0b80-1d20-41ce-a785-b33efd6af3a5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.620358] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa9e6f71-e7c0-46d6-8a89-5d3db0f2aad2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.635007] env[61243]: DEBUG nova.compute.provider_tree [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.876144] env[61243]: ERROR nova.compute.manager [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e148ba07-8dcb-4345-93fa-15a43fbbbe3f, please check neutron logs for more information. [ 648.876144] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 648.876144] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.876144] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 648.876144] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.876144] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 648.876144] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.876144] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 648.876144] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.876144] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 648.876144] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.876144] env[61243]: ERROR nova.compute.manager raise self.value [ 648.876144] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.876144] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 648.876144] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.876144] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 648.877540] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.877540] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 648.877540] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e148ba07-8dcb-4345-93fa-15a43fbbbe3f, please check neutron logs for more information. [ 648.877540] env[61243]: ERROR nova.compute.manager [ 648.877540] env[61243]: Traceback (most recent call last): [ 648.877540] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 648.877540] env[61243]: listener.cb(fileno) [ 648.877540] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.877540] env[61243]: result = function(*args, **kwargs) [ 648.877540] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 648.877540] env[61243]: return func(*args, **kwargs) [ 648.877540] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.877540] env[61243]: raise e [ 648.877540] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.877540] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 648.877540] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.877540] env[61243]: created_port_ids = self._update_ports_for_instance( [ 648.877540] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.877540] env[61243]: with excutils.save_and_reraise_exception(): [ 648.877540] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.877540] env[61243]: self.force_reraise() [ 648.877540] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.877540] env[61243]: raise self.value [ 648.877540] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.877540] env[61243]: updated_port = self._update_port( [ 648.877540] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.877540] env[61243]: _ensure_no_port_binding_failure(port) [ 648.877540] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.877540] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 648.879258] env[61243]: nova.exception.PortBindingFailed: Binding failed for port e148ba07-8dcb-4345-93fa-15a43fbbbe3f, please check neutron logs for more information. [ 648.879258] env[61243]: Removing descriptor: 16 [ 649.019436] env[61243]: DEBUG nova.network.neutron [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.096066] env[61243]: DEBUG nova.network.neutron [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.138714] env[61243]: DEBUG nova.scheduler.client.report [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.211709] env[61243]: DEBUG nova.compute.manager [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 649.236576] env[61243]: DEBUG nova.virt.hardware [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 649.236812] env[61243]: DEBUG nova.virt.hardware [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 649.236968] env[61243]: DEBUG nova.virt.hardware [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 649.237163] env[61243]: DEBUG nova.virt.hardware [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 649.237314] env[61243]: DEBUG nova.virt.hardware [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 649.237457] env[61243]: DEBUG nova.virt.hardware [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 649.237667] env[61243]: DEBUG nova.virt.hardware [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 649.237827] env[61243]: DEBUG nova.virt.hardware [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 649.237993] env[61243]: DEBUG nova.virt.hardware [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 649.238175] env[61243]: DEBUG nova.virt.hardware [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 649.238345] env[61243]: DEBUG nova.virt.hardware [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 649.239193] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69bf663b-93a6-4052-af46-05a9006c2b4e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.247327] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaa8da85-8b25-45c7-894f-75a42a34bf4d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.260575] env[61243]: ERROR nova.compute.manager [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e148ba07-8dcb-4345-93fa-15a43fbbbe3f, please check neutron logs for more information. [ 649.260575] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Traceback (most recent call last): [ 649.260575] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 649.260575] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] yield resources [ 649.260575] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 649.260575] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] self.driver.spawn(context, instance, image_meta, [ 649.260575] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 649.260575] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 649.260575] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 649.260575] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] vm_ref = self.build_virtual_machine(instance, [ 649.260575] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 649.261020] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] vif_infos = vmwarevif.get_vif_info(self._session, [ 649.261020] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 649.261020] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] for vif in network_info: [ 649.261020] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 649.261020] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] return self._sync_wrapper(fn, *args, **kwargs) [ 649.261020] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 649.261020] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] self.wait() [ 649.261020] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 649.261020] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] self[:] = self._gt.wait() [ 649.261020] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 649.261020] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] return self._exit_event.wait() [ 649.261020] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 649.261020] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] current.throw(*self._exc) [ 649.261467] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.261467] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] result = function(*args, **kwargs) [ 649.261467] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 649.261467] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] return func(*args, **kwargs) [ 649.261467] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 649.261467] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] raise e [ 649.261467] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.261467] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] nwinfo = self.network_api.allocate_for_instance( [ 649.261467] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.261467] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] created_port_ids = self._update_ports_for_instance( [ 649.261467] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.261467] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] with excutils.save_and_reraise_exception(): [ 649.261467] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.261841] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] self.force_reraise() [ 649.261841] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.261841] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] raise self.value [ 649.261841] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.261841] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] updated_port = self._update_port( [ 649.261841] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.261841] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] _ensure_no_port_binding_failure(port) [ 649.261841] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.261841] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] raise exception.PortBindingFailed(port_id=port['id']) [ 649.261841] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] nova.exception.PortBindingFailed: Binding failed for port e148ba07-8dcb-4345-93fa-15a43fbbbe3f, please check neutron logs for more information. [ 649.261841] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] [ 649.261841] env[61243]: INFO nova.compute.manager [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Terminating instance [ 649.262921] env[61243]: DEBUG oslo_concurrency.lockutils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Acquiring lock "refresh_cache-64efd676-9166-4ffa-a437-6fa6c466ace0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.263108] env[61243]: DEBUG oslo_concurrency.lockutils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Acquired lock "refresh_cache-64efd676-9166-4ffa-a437-6fa6c466ace0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.263284] env[61243]: DEBUG nova.network.neutron [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 649.599109] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Releasing lock "refresh_cache-0625ac10-55b4-488a-aba5-bdaa394045f4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.599442] env[61243]: DEBUG nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 649.599859] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eb2158c6-43d9-4a5b-acf5-10ad64b51e17 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.608635] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b9d119-1994-491d-a18c-88b22ae7490f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.629270] env[61243]: WARNING nova.virt.vmwareapi.driver [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 0625ac10-55b4-488a-aba5-bdaa394045f4 could not be found. [ 649.629515] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 649.629800] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-95545839-f4d5-4e6e-9677-079f9633ab65 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.636971] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a695d4d9-bef1-475c-ba30-56cc656a0dbb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.648468] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.648943] env[61243]: DEBUG nova.compute.manager [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 649.651579] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.049s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.653054] env[61243]: INFO nova.compute.claims [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 649.665793] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0625ac10-55b4-488a-aba5-bdaa394045f4 could not be found. [ 649.666018] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 649.666204] env[61243]: INFO nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Took 0.07 seconds to destroy the instance on the hypervisor. [ 649.666445] env[61243]: DEBUG oslo.service.loopingcall [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 649.666704] env[61243]: DEBUG nova.compute.manager [-] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 649.666754] env[61243]: DEBUG nova.network.neutron [-] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 649.682443] env[61243]: DEBUG nova.network.neutron [-] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.783086] env[61243]: DEBUG nova.network.neutron [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.862945] env[61243]: DEBUG nova.network.neutron [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.158021] env[61243]: DEBUG nova.compute.utils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 650.161205] env[61243]: DEBUG nova.compute.manager [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 650.161380] env[61243]: DEBUG nova.network.neutron [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 650.186287] env[61243]: DEBUG nova.network.neutron [-] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.200066] env[61243]: DEBUG nova.policy [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '28d7b73f55f743a1a7126cc85613ebb7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d983b7bb0224a50a26b0334e697422f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 650.261925] env[61243]: DEBUG nova.compute.manager [req-94c62550-b242-4963-a71b-fa106a68be12 req-41237394-ebc1-48fc-87b4-79fd26e7ba5a service nova] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Received event network-changed-e148ba07-8dcb-4345-93fa-15a43fbbbe3f {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 650.262143] env[61243]: DEBUG nova.compute.manager [req-94c62550-b242-4963-a71b-fa106a68be12 req-41237394-ebc1-48fc-87b4-79fd26e7ba5a service nova] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Refreshing instance network info cache due to event network-changed-e148ba07-8dcb-4345-93fa-15a43fbbbe3f. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 650.262345] env[61243]: DEBUG oslo_concurrency.lockutils [req-94c62550-b242-4963-a71b-fa106a68be12 req-41237394-ebc1-48fc-87b4-79fd26e7ba5a service nova] Acquiring lock "refresh_cache-64efd676-9166-4ffa-a437-6fa6c466ace0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.366366] env[61243]: DEBUG oslo_concurrency.lockutils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Releasing lock "refresh_cache-64efd676-9166-4ffa-a437-6fa6c466ace0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.366480] env[61243]: DEBUG nova.compute.manager [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 650.366658] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 650.367875] env[61243]: DEBUG oslo_concurrency.lockutils [req-94c62550-b242-4963-a71b-fa106a68be12 req-41237394-ebc1-48fc-87b4-79fd26e7ba5a service nova] Acquired lock "refresh_cache-64efd676-9166-4ffa-a437-6fa6c466ace0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.367875] env[61243]: DEBUG nova.network.neutron [req-94c62550-b242-4963-a71b-fa106a68be12 req-41237394-ebc1-48fc-87b4-79fd26e7ba5a service nova] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Refreshing network info cache for port e148ba07-8dcb-4345-93fa-15a43fbbbe3f {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 650.372025] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d408761a-e056-4b6c-8173-232493f3e7be {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.379157] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d43fe4f6-520b-4c20-b58d-da8f4748a18a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.402892] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 64efd676-9166-4ffa-a437-6fa6c466ace0 could not be found. [ 650.403580] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 650.403580] env[61243]: INFO nova.compute.manager [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 650.403708] env[61243]: DEBUG oslo.service.loopingcall [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 650.404333] env[61243]: DEBUG nova.compute.manager [-] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 650.404333] env[61243]: DEBUG nova.network.neutron [-] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 650.422509] env[61243]: DEBUG nova.network.neutron [-] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.488894] env[61243]: DEBUG nova.network.neutron [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Successfully created port: a073df5c-76be-4a85-b48a-b7a65a1d7482 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 650.661746] env[61243]: DEBUG nova.compute.manager [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 650.690666] env[61243]: INFO nova.compute.manager [-] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Took 1.02 seconds to deallocate network for instance. [ 650.896906] env[61243]: DEBUG nova.network.neutron [req-94c62550-b242-4963-a71b-fa106a68be12 req-41237394-ebc1-48fc-87b4-79fd26e7ba5a service nova] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.926384] env[61243]: DEBUG nova.network.neutron [-] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.013806] env[61243]: DEBUG nova.network.neutron [req-94c62550-b242-4963-a71b-fa106a68be12 req-41237394-ebc1-48fc-87b4-79fd26e7ba5a service nova] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.074261] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30934fc-1946-4163-aedf-234218ed6e9f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.082325] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49a71d48-f746-4c1d-a841-995d50389212 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.111952] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c875495-c731-431e-ae13-a9e6740f947d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.119175] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7fca5ec-c0d1-45c6-a242-028541d8b05e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.132359] env[61243]: DEBUG nova.compute.provider_tree [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.252634] env[61243]: INFO nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Took 0.56 seconds to detach 1 volumes for instance. [ 651.254773] env[61243]: DEBUG nova.compute.claims [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 651.254966] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.430572] env[61243]: ERROR nova.compute.manager [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a073df5c-76be-4a85-b48a-b7a65a1d7482, please check neutron logs for more information. [ 651.430572] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 651.430572] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.430572] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 651.430572] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.430572] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 651.430572] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.430572] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 651.430572] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.430572] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 651.430572] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.430572] env[61243]: ERROR nova.compute.manager raise self.value [ 651.430572] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.430572] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 651.430572] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.430572] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 651.431046] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.431046] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 651.431046] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a073df5c-76be-4a85-b48a-b7a65a1d7482, please check neutron logs for more information. [ 651.431046] env[61243]: ERROR nova.compute.manager [ 651.431046] env[61243]: Traceback (most recent call last): [ 651.431046] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 651.431046] env[61243]: listener.cb(fileno) [ 651.431046] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.431046] env[61243]: result = function(*args, **kwargs) [ 651.431046] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 651.431046] env[61243]: return func(*args, **kwargs) [ 651.431046] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.431046] env[61243]: raise e [ 651.431046] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.431046] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 651.431046] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.431046] env[61243]: created_port_ids = self._update_ports_for_instance( [ 651.431046] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.431046] env[61243]: with excutils.save_and_reraise_exception(): [ 651.431046] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.431046] env[61243]: self.force_reraise() [ 651.431046] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.431046] env[61243]: raise self.value [ 651.431046] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.431046] env[61243]: updated_port = self._update_port( [ 651.431046] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.431046] env[61243]: _ensure_no_port_binding_failure(port) [ 651.431046] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.431046] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 651.431746] env[61243]: nova.exception.PortBindingFailed: Binding failed for port a073df5c-76be-4a85-b48a-b7a65a1d7482, please check neutron logs for more information. [ 651.431746] env[61243]: Removing descriptor: 16 [ 651.431746] env[61243]: INFO nova.compute.manager [-] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Took 1.03 seconds to deallocate network for instance. [ 651.433446] env[61243]: DEBUG nova.compute.claims [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 651.433638] env[61243]: DEBUG oslo_concurrency.lockutils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.518536] env[61243]: DEBUG oslo_concurrency.lockutils [req-94c62550-b242-4963-a71b-fa106a68be12 req-41237394-ebc1-48fc-87b4-79fd26e7ba5a service nova] Releasing lock "refresh_cache-64efd676-9166-4ffa-a437-6fa6c466ace0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.518809] env[61243]: DEBUG nova.compute.manager [req-94c62550-b242-4963-a71b-fa106a68be12 req-41237394-ebc1-48fc-87b4-79fd26e7ba5a service nova] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Received event network-vif-deleted-e148ba07-8dcb-4345-93fa-15a43fbbbe3f {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 651.635240] env[61243]: DEBUG nova.scheduler.client.report [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 651.677953] env[61243]: DEBUG nova.compute.manager [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 651.702670] env[61243]: DEBUG nova.virt.hardware [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 651.702908] env[61243]: DEBUG nova.virt.hardware [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 651.703080] env[61243]: DEBUG nova.virt.hardware [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 651.703266] env[61243]: DEBUG nova.virt.hardware [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 651.703415] env[61243]: DEBUG nova.virt.hardware [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 651.703560] env[61243]: DEBUG nova.virt.hardware [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 651.703762] env[61243]: DEBUG nova.virt.hardware [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 651.703921] env[61243]: DEBUG nova.virt.hardware [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 651.704097] env[61243]: DEBUG nova.virt.hardware [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 651.704265] env[61243]: DEBUG nova.virt.hardware [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 651.704463] env[61243]: DEBUG nova.virt.hardware [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 651.705294] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c718ad-ba68-466d-aa43-c370fd973bb8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.713101] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3964864f-f6ce-44f2-a227-e77f3d1a21ea {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.727990] env[61243]: ERROR nova.compute.manager [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a073df5c-76be-4a85-b48a-b7a65a1d7482, please check neutron logs for more information. [ 651.727990] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Traceback (most recent call last): [ 651.727990] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 651.727990] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] yield resources [ 651.727990] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 651.727990] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] self.driver.spawn(context, instance, image_meta, [ 651.727990] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 651.727990] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.727990] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.727990] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] vm_ref = self.build_virtual_machine(instance, [ 651.727990] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.728424] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.728424] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.728424] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] for vif in network_info: [ 651.728424] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 651.728424] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] return self._sync_wrapper(fn, *args, **kwargs) [ 651.728424] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 651.728424] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] self.wait() [ 651.728424] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 651.728424] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] self[:] = self._gt.wait() [ 651.728424] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.728424] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] return self._exit_event.wait() [ 651.728424] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 651.728424] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] current.throw(*self._exc) [ 651.728802] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.728802] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] result = function(*args, **kwargs) [ 651.728802] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 651.728802] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] return func(*args, **kwargs) [ 651.728802] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.728802] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] raise e [ 651.728802] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.728802] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] nwinfo = self.network_api.allocate_for_instance( [ 651.728802] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.728802] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] created_port_ids = self._update_ports_for_instance( [ 651.728802] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.728802] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] with excutils.save_and_reraise_exception(): [ 651.728802] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.729179] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] self.force_reraise() [ 651.729179] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.729179] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] raise self.value [ 651.729179] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.729179] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] updated_port = self._update_port( [ 651.729179] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.729179] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] _ensure_no_port_binding_failure(port) [ 651.729179] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.729179] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] raise exception.PortBindingFailed(port_id=port['id']) [ 651.729179] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] nova.exception.PortBindingFailed: Binding failed for port a073df5c-76be-4a85-b48a-b7a65a1d7482, please check neutron logs for more information. [ 651.729179] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] [ 651.729179] env[61243]: INFO nova.compute.manager [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Terminating instance [ 651.730448] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Acquiring lock "refresh_cache-8efd391b-7b76-4167-bd18-48566271ccc8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.730613] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Acquired lock "refresh_cache-8efd391b-7b76-4167-bd18-48566271ccc8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.730777] env[61243]: DEBUG nova.network.neutron [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 652.140538] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.489s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.141085] env[61243]: DEBUG nova.compute.manager [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 652.144036] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.244s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.146205] env[61243]: INFO nova.compute.claims [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.248731] env[61243]: DEBUG nova.network.neutron [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.291023] env[61243]: DEBUG nova.compute.manager [req-c64f9fdd-0de8-4b70-992c-ff3cbd52ea5c req-5bf8e805-fce4-47ce-ab66-55a78102f325 service nova] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Received event network-changed-a073df5c-76be-4a85-b48a-b7a65a1d7482 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 652.291226] env[61243]: DEBUG nova.compute.manager [req-c64f9fdd-0de8-4b70-992c-ff3cbd52ea5c req-5bf8e805-fce4-47ce-ab66-55a78102f325 service nova] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Refreshing instance network info cache due to event network-changed-a073df5c-76be-4a85-b48a-b7a65a1d7482. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 652.291462] env[61243]: DEBUG oslo_concurrency.lockutils [req-c64f9fdd-0de8-4b70-992c-ff3cbd52ea5c req-5bf8e805-fce4-47ce-ab66-55a78102f325 service nova] Acquiring lock "refresh_cache-8efd391b-7b76-4167-bd18-48566271ccc8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.331540] env[61243]: DEBUG nova.network.neutron [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.650352] env[61243]: DEBUG nova.compute.utils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 652.654170] env[61243]: DEBUG nova.compute.manager [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 652.654349] env[61243]: DEBUG nova.network.neutron [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 652.700743] env[61243]: DEBUG nova.policy [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '982ea79b687d4533b29627ae919c1064', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f5fc6d2d9e24f85bb18c6643d9ebe5d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 652.834558] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Releasing lock "refresh_cache-8efd391b-7b76-4167-bd18-48566271ccc8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.835052] env[61243]: DEBUG nova.compute.manager [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 652.835298] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 652.835638] env[61243]: DEBUG oslo_concurrency.lockutils [req-c64f9fdd-0de8-4b70-992c-ff3cbd52ea5c req-5bf8e805-fce4-47ce-ab66-55a78102f325 service nova] Acquired lock "refresh_cache-8efd391b-7b76-4167-bd18-48566271ccc8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.835801] env[61243]: DEBUG nova.network.neutron [req-c64f9fdd-0de8-4b70-992c-ff3cbd52ea5c req-5bf8e805-fce4-47ce-ab66-55a78102f325 service nova] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Refreshing network info cache for port a073df5c-76be-4a85-b48a-b7a65a1d7482 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 652.837247] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-98cc205a-8a34-46f7-b75c-b6d3f1779fa4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.847328] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54562665-7640-4f18-a6b5-dd12044c476a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.870926] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8efd391b-7b76-4167-bd18-48566271ccc8 could not be found. [ 652.871182] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 652.871373] env[61243]: INFO nova.compute.manager [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 652.871677] env[61243]: DEBUG oslo.service.loopingcall [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 652.872192] env[61243]: DEBUG nova.compute.manager [-] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 652.872192] env[61243]: DEBUG nova.network.neutron [-] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 652.891741] env[61243]: DEBUG nova.network.neutron [-] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.004580] env[61243]: DEBUG nova.network.neutron [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Successfully created port: 6eb50563-62d3-4e2a-8adf-9d0d5417e4f7 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 653.157788] env[61243]: DEBUG nova.compute.manager [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 653.360171] env[61243]: DEBUG nova.network.neutron [req-c64f9fdd-0de8-4b70-992c-ff3cbd52ea5c req-5bf8e805-fce4-47ce-ab66-55a78102f325 service nova] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.398349] env[61243]: DEBUG nova.network.neutron [-] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.476578] env[61243]: DEBUG nova.network.neutron [req-c64f9fdd-0de8-4b70-992c-ff3cbd52ea5c req-5bf8e805-fce4-47ce-ab66-55a78102f325 service nova] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.558181] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-853710f7-c0d9-4143-ad4b-753e1817a4f1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.566069] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77766352-6f03-4afd-af7c-df9a07177db2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.598214] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0a5e9c-f314-40e3-89d0-5a8e1ad47753 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.606642] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ba0acd-3366-44e9-b85d-5594662af063 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.620160] env[61243]: DEBUG nova.compute.provider_tree [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.902075] env[61243]: INFO nova.compute.manager [-] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Took 1.03 seconds to deallocate network for instance. [ 653.904659] env[61243]: ERROR nova.compute.manager [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6eb50563-62d3-4e2a-8adf-9d0d5417e4f7, please check neutron logs for more information. [ 653.904659] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 653.904659] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.904659] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 653.904659] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.904659] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 653.904659] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.904659] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 653.904659] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.904659] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 653.904659] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.904659] env[61243]: ERROR nova.compute.manager raise self.value [ 653.904659] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.904659] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 653.904659] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.904659] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 653.905213] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.905213] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 653.905213] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6eb50563-62d3-4e2a-8adf-9d0d5417e4f7, please check neutron logs for more information. [ 653.905213] env[61243]: ERROR nova.compute.manager [ 653.905213] env[61243]: Traceback (most recent call last): [ 653.905213] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 653.905213] env[61243]: listener.cb(fileno) [ 653.905213] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.905213] env[61243]: result = function(*args, **kwargs) [ 653.905213] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 653.905213] env[61243]: return func(*args, **kwargs) [ 653.905213] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.905213] env[61243]: raise e [ 653.905213] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.905213] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 653.905213] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.905213] env[61243]: created_port_ids = self._update_ports_for_instance( [ 653.905213] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.905213] env[61243]: with excutils.save_and_reraise_exception(): [ 653.905213] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.905213] env[61243]: self.force_reraise() [ 653.905213] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.905213] env[61243]: raise self.value [ 653.905213] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.905213] env[61243]: updated_port = self._update_port( [ 653.905213] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.905213] env[61243]: _ensure_no_port_binding_failure(port) [ 653.905213] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.905213] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 653.906059] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 6eb50563-62d3-4e2a-8adf-9d0d5417e4f7, please check neutron logs for more information. [ 653.906059] env[61243]: Removing descriptor: 16 [ 653.906059] env[61243]: DEBUG nova.compute.claims [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 653.906059] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.979775] env[61243]: DEBUG oslo_concurrency.lockutils [req-c64f9fdd-0de8-4b70-992c-ff3cbd52ea5c req-5bf8e805-fce4-47ce-ab66-55a78102f325 service nova] Releasing lock "refresh_cache-8efd391b-7b76-4167-bd18-48566271ccc8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.980027] env[61243]: DEBUG nova.compute.manager [req-c64f9fdd-0de8-4b70-992c-ff3cbd52ea5c req-5bf8e805-fce4-47ce-ab66-55a78102f325 service nova] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Received event network-vif-deleted-a073df5c-76be-4a85-b48a-b7a65a1d7482 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 654.123474] env[61243]: DEBUG nova.scheduler.client.report [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.167624] env[61243]: DEBUG nova.compute.manager [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 654.193480] env[61243]: DEBUG nova.virt.hardware [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 654.193717] env[61243]: DEBUG nova.virt.hardware [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 654.193875] env[61243]: DEBUG nova.virt.hardware [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 654.194070] env[61243]: DEBUG nova.virt.hardware [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 654.194224] env[61243]: DEBUG nova.virt.hardware [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 654.194401] env[61243]: DEBUG nova.virt.hardware [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 654.194624] env[61243]: DEBUG nova.virt.hardware [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 654.194784] env[61243]: DEBUG nova.virt.hardware [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 654.194949] env[61243]: DEBUG nova.virt.hardware [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 654.195125] env[61243]: DEBUG nova.virt.hardware [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 654.195303] env[61243]: DEBUG nova.virt.hardware [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 654.196182] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751e2116-ffcb-4f00-947f-db5f898532d9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.203876] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-508319b9-2bc0-4f4e-baad-062d26f196f7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.217935] env[61243]: ERROR nova.compute.manager [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6eb50563-62d3-4e2a-8adf-9d0d5417e4f7, please check neutron logs for more information. [ 654.217935] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Traceback (most recent call last): [ 654.217935] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 654.217935] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] yield resources [ 654.217935] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 654.217935] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] self.driver.spawn(context, instance, image_meta, [ 654.217935] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 654.217935] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.217935] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.217935] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] vm_ref = self.build_virtual_machine(instance, [ 654.217935] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.218370] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.218370] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.218370] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] for vif in network_info: [ 654.218370] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.218370] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] return self._sync_wrapper(fn, *args, **kwargs) [ 654.218370] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.218370] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] self.wait() [ 654.218370] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.218370] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] self[:] = self._gt.wait() [ 654.218370] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.218370] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] return self._exit_event.wait() [ 654.218370] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 654.218370] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] current.throw(*self._exc) [ 654.218808] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.218808] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] result = function(*args, **kwargs) [ 654.218808] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 654.218808] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] return func(*args, **kwargs) [ 654.218808] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.218808] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] raise e [ 654.218808] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.218808] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] nwinfo = self.network_api.allocate_for_instance( [ 654.218808] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.218808] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] created_port_ids = self._update_ports_for_instance( [ 654.218808] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.218808] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] with excutils.save_and_reraise_exception(): [ 654.218808] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.219192] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] self.force_reraise() [ 654.219192] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.219192] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] raise self.value [ 654.219192] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.219192] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] updated_port = self._update_port( [ 654.219192] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.219192] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] _ensure_no_port_binding_failure(port) [ 654.219192] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.219192] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] raise exception.PortBindingFailed(port_id=port['id']) [ 654.219192] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] nova.exception.PortBindingFailed: Binding failed for port 6eb50563-62d3-4e2a-8adf-9d0d5417e4f7, please check neutron logs for more information. [ 654.219192] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] [ 654.219192] env[61243]: INFO nova.compute.manager [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Terminating instance [ 654.220311] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Acquiring lock "refresh_cache-84529438-8560-4323-a2f8-86cb917d18e6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.220468] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Acquired lock "refresh_cache-84529438-8560-4323-a2f8-86cb917d18e6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.220640] env[61243]: DEBUG nova.network.neutron [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 654.344856] env[61243]: DEBUG nova.compute.manager [req-e51a62cd-4077-405b-aab7-41150166e945 req-24d31ea8-49ab-4115-b6c7-deeb6d71398a service nova] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Received event network-changed-6eb50563-62d3-4e2a-8adf-9d0d5417e4f7 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 654.345149] env[61243]: DEBUG nova.compute.manager [req-e51a62cd-4077-405b-aab7-41150166e945 req-24d31ea8-49ab-4115-b6c7-deeb6d71398a service nova] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Refreshing instance network info cache due to event network-changed-6eb50563-62d3-4e2a-8adf-9d0d5417e4f7. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 654.345375] env[61243]: DEBUG oslo_concurrency.lockutils [req-e51a62cd-4077-405b-aab7-41150166e945 req-24d31ea8-49ab-4115-b6c7-deeb6d71398a service nova] Acquiring lock "refresh_cache-84529438-8560-4323-a2f8-86cb917d18e6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.628552] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.629197] env[61243]: DEBUG nova.compute.manager [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 654.633644] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.811s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.635048] env[61243]: INFO nova.compute.claims [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 654.740676] env[61243]: DEBUG nova.network.neutron [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.795813] env[61243]: DEBUG nova.network.neutron [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.138930] env[61243]: DEBUG nova.compute.utils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 655.143286] env[61243]: DEBUG nova.compute.manager [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 655.143497] env[61243]: DEBUG nova.network.neutron [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 655.190764] env[61243]: DEBUG nova.policy [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '982ea79b687d4533b29627ae919c1064', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f5fc6d2d9e24f85bb18c6643d9ebe5d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 655.299092] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Releasing lock "refresh_cache-84529438-8560-4323-a2f8-86cb917d18e6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.299092] env[61243]: DEBUG nova.compute.manager [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 655.299092] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 655.299345] env[61243]: DEBUG oslo_concurrency.lockutils [req-e51a62cd-4077-405b-aab7-41150166e945 req-24d31ea8-49ab-4115-b6c7-deeb6d71398a service nova] Acquired lock "refresh_cache-84529438-8560-4323-a2f8-86cb917d18e6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.299513] env[61243]: DEBUG nova.network.neutron [req-e51a62cd-4077-405b-aab7-41150166e945 req-24d31ea8-49ab-4115-b6c7-deeb6d71398a service nova] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Refreshing network info cache for port 6eb50563-62d3-4e2a-8adf-9d0d5417e4f7 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 655.300601] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-90f50bc7-d2d0-4ded-9897-8a5c803782e1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.309644] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a7b703-2287-4927-9e19-673abbadcd0f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.332038] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 84529438-8560-4323-a2f8-86cb917d18e6 could not be found. [ 655.332257] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 655.332433] env[61243]: INFO nova.compute.manager [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 655.332663] env[61243]: DEBUG oslo.service.loopingcall [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.332921] env[61243]: DEBUG nova.compute.manager [-] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.332995] env[61243]: DEBUG nova.network.neutron [-] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 655.353053] env[61243]: DEBUG nova.network.neutron [-] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.456368] env[61243]: DEBUG nova.network.neutron [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Successfully created port: c1b660a5-8cfb-4650-9a8b-fa7bf0c491ab {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 655.646848] env[61243]: DEBUG nova.compute.manager [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 655.824504] env[61243]: DEBUG nova.network.neutron [req-e51a62cd-4077-405b-aab7-41150166e945 req-24d31ea8-49ab-4115-b6c7-deeb6d71398a service nova] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.855600] env[61243]: DEBUG nova.network.neutron [-] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.033996] env[61243]: DEBUG nova.network.neutron [req-e51a62cd-4077-405b-aab7-41150166e945 req-24d31ea8-49ab-4115-b6c7-deeb6d71398a service nova] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.066627] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb4ccdcb-7eee-4ee4-9298-7b26a092b18a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.074957] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3068050a-1f85-489c-81c8-3ec6f0630b57 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.109457] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0900b59b-d9e4-4681-be30-09410d17205b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.118335] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f9b68eb-eff5-4e3c-803a-0d0cec3c04ea {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.132783] env[61243]: DEBUG nova.compute.provider_tree [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.359345] env[61243]: INFO nova.compute.manager [-] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Took 1.03 seconds to deallocate network for instance. [ 656.362540] env[61243]: DEBUG nova.compute.claims [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 656.362749] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.378607] env[61243]: DEBUG nova.compute.manager [req-394359a0-91da-4d62-8872-e6a1dd23ec0c req-df0db2ee-e908-4e91-a37a-29d1f0ef25b9 service nova] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Received event network-changed-c1b660a5-8cfb-4650-9a8b-fa7bf0c491ab {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 656.378607] env[61243]: DEBUG nova.compute.manager [req-394359a0-91da-4d62-8872-e6a1dd23ec0c req-df0db2ee-e908-4e91-a37a-29d1f0ef25b9 service nova] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Refreshing instance network info cache due to event network-changed-c1b660a5-8cfb-4650-9a8b-fa7bf0c491ab. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 656.379113] env[61243]: DEBUG oslo_concurrency.lockutils [req-394359a0-91da-4d62-8872-e6a1dd23ec0c req-df0db2ee-e908-4e91-a37a-29d1f0ef25b9 service nova] Acquiring lock "refresh_cache-4072d5a6-f69d-4fb9-837b-7855054e5206" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.379113] env[61243]: DEBUG oslo_concurrency.lockutils [req-394359a0-91da-4d62-8872-e6a1dd23ec0c req-df0db2ee-e908-4e91-a37a-29d1f0ef25b9 service nova] Acquired lock "refresh_cache-4072d5a6-f69d-4fb9-837b-7855054e5206" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.379184] env[61243]: DEBUG nova.network.neutron [req-394359a0-91da-4d62-8872-e6a1dd23ec0c req-df0db2ee-e908-4e91-a37a-29d1f0ef25b9 service nova] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Refreshing network info cache for port c1b660a5-8cfb-4650-9a8b-fa7bf0c491ab {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 656.441103] env[61243]: ERROR nova.compute.manager [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c1b660a5-8cfb-4650-9a8b-fa7bf0c491ab, please check neutron logs for more information. [ 656.441103] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 656.441103] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.441103] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 656.441103] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.441103] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 656.441103] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.441103] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 656.441103] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.441103] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 656.441103] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.441103] env[61243]: ERROR nova.compute.manager raise self.value [ 656.441103] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.441103] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 656.441103] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.441103] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 656.441536] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.441536] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 656.441536] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c1b660a5-8cfb-4650-9a8b-fa7bf0c491ab, please check neutron logs for more information. [ 656.441536] env[61243]: ERROR nova.compute.manager [ 656.441536] env[61243]: Traceback (most recent call last): [ 656.441536] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 656.441536] env[61243]: listener.cb(fileno) [ 656.441536] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.441536] env[61243]: result = function(*args, **kwargs) [ 656.441536] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.441536] env[61243]: return func(*args, **kwargs) [ 656.441536] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.441536] env[61243]: raise e [ 656.441536] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.441536] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 656.441536] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.441536] env[61243]: created_port_ids = self._update_ports_for_instance( [ 656.441536] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.441536] env[61243]: with excutils.save_and_reraise_exception(): [ 656.441536] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.441536] env[61243]: self.force_reraise() [ 656.441536] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.441536] env[61243]: raise self.value [ 656.441536] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.441536] env[61243]: updated_port = self._update_port( [ 656.441536] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.441536] env[61243]: _ensure_no_port_binding_failure(port) [ 656.441536] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.441536] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 656.442698] env[61243]: nova.exception.PortBindingFailed: Binding failed for port c1b660a5-8cfb-4650-9a8b-fa7bf0c491ab, please check neutron logs for more information. [ 656.442698] env[61243]: Removing descriptor: 16 [ 656.539037] env[61243]: DEBUG oslo_concurrency.lockutils [req-e51a62cd-4077-405b-aab7-41150166e945 req-24d31ea8-49ab-4115-b6c7-deeb6d71398a service nova] Releasing lock "refresh_cache-84529438-8560-4323-a2f8-86cb917d18e6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.539321] env[61243]: DEBUG nova.compute.manager [req-e51a62cd-4077-405b-aab7-41150166e945 req-24d31ea8-49ab-4115-b6c7-deeb6d71398a service nova] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Received event network-vif-deleted-6eb50563-62d3-4e2a-8adf-9d0d5417e4f7 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 656.636672] env[61243]: DEBUG nova.scheduler.client.report [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.658229] env[61243]: DEBUG nova.compute.manager [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 656.682066] env[61243]: DEBUG nova.virt.hardware [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 656.682293] env[61243]: DEBUG nova.virt.hardware [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 656.682454] env[61243]: DEBUG nova.virt.hardware [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 656.682638] env[61243]: DEBUG nova.virt.hardware [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 656.682807] env[61243]: DEBUG nova.virt.hardware [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 656.682956] env[61243]: DEBUG nova.virt.hardware [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 656.683179] env[61243]: DEBUG nova.virt.hardware [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 656.683340] env[61243]: DEBUG nova.virt.hardware [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 656.683503] env[61243]: DEBUG nova.virt.hardware [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 656.683665] env[61243]: DEBUG nova.virt.hardware [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 656.683833] env[61243]: DEBUG nova.virt.hardware [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 656.684700] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5756346f-4d70-4597-aeaa-fda88b3fddee {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.692754] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-223b323e-5873-4682-8e9c-4aefb1bc5c63 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.706198] env[61243]: ERROR nova.compute.manager [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c1b660a5-8cfb-4650-9a8b-fa7bf0c491ab, please check neutron logs for more information. [ 656.706198] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Traceback (most recent call last): [ 656.706198] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 656.706198] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] yield resources [ 656.706198] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.706198] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] self.driver.spawn(context, instance, image_meta, [ 656.706198] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 656.706198] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.706198] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.706198] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] vm_ref = self.build_virtual_machine(instance, [ 656.706198] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.706565] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.706565] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.706565] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] for vif in network_info: [ 656.706565] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.706565] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] return self._sync_wrapper(fn, *args, **kwargs) [ 656.706565] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.706565] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] self.wait() [ 656.706565] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.706565] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] self[:] = self._gt.wait() [ 656.706565] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.706565] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] return self._exit_event.wait() [ 656.706565] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 656.706565] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] current.throw(*self._exc) [ 656.706925] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.706925] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] result = function(*args, **kwargs) [ 656.706925] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.706925] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] return func(*args, **kwargs) [ 656.706925] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.706925] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] raise e [ 656.706925] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.706925] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] nwinfo = self.network_api.allocate_for_instance( [ 656.706925] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.706925] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] created_port_ids = self._update_ports_for_instance( [ 656.706925] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.706925] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] with excutils.save_and_reraise_exception(): [ 656.706925] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.707315] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] self.force_reraise() [ 656.707315] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.707315] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] raise self.value [ 656.707315] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.707315] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] updated_port = self._update_port( [ 656.707315] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.707315] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] _ensure_no_port_binding_failure(port) [ 656.707315] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.707315] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] raise exception.PortBindingFailed(port_id=port['id']) [ 656.707315] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] nova.exception.PortBindingFailed: Binding failed for port c1b660a5-8cfb-4650-9a8b-fa7bf0c491ab, please check neutron logs for more information. [ 656.707315] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] [ 656.707315] env[61243]: INFO nova.compute.manager [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Terminating instance [ 656.708418] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Acquiring lock "refresh_cache-4072d5a6-f69d-4fb9-837b-7855054e5206" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.897038] env[61243]: DEBUG nova.network.neutron [req-394359a0-91da-4d62-8872-e6a1dd23ec0c req-df0db2ee-e908-4e91-a37a-29d1f0ef25b9 service nova] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.954080] env[61243]: DEBUG nova.network.neutron [req-394359a0-91da-4d62-8872-e6a1dd23ec0c req-df0db2ee-e908-4e91-a37a-29d1f0ef25b9 service nova] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.141694] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.508s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.142502] env[61243]: DEBUG nova.compute.manager [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 657.145369] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.216s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.457251] env[61243]: DEBUG oslo_concurrency.lockutils [req-394359a0-91da-4d62-8872-e6a1dd23ec0c req-df0db2ee-e908-4e91-a37a-29d1f0ef25b9 service nova] Releasing lock "refresh_cache-4072d5a6-f69d-4fb9-837b-7855054e5206" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.457806] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Acquired lock "refresh_cache-4072d5a6-f69d-4fb9-837b-7855054e5206" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.458095] env[61243]: DEBUG nova.network.neutron [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 657.646955] env[61243]: DEBUG nova.compute.utils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 657.648415] env[61243]: DEBUG nova.compute.manager [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 657.648600] env[61243]: DEBUG nova.network.neutron [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 657.692124] env[61243]: DEBUG nova.policy [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '29275dd6899d4982bcacbdb35ae2de9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e697209eb99e4199891657e359f32690', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 657.949515] env[61243]: DEBUG nova.network.neutron [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Successfully created port: 0d346e84-b8e6-4678-99ad-bf03d7a94af8 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 657.996069] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458502fb-e4bb-49f9-81c7-2c21b1162c32 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.003861] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c8b0df0-3765-4dc8-92c6-44ca9ad7c885 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.007349] env[61243]: DEBUG nova.network.neutron [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.036885] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd0f3fc-63c1-4412-9d0a-a783c28cd85e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.044125] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-220bee79-72b7-45dc-b588-3826f79791bd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.057474] env[61243]: DEBUG nova.compute.provider_tree [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.099929] env[61243]: DEBUG nova.network.neutron [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.151664] env[61243]: DEBUG nova.compute.manager [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 658.405310] env[61243]: DEBUG nova.compute.manager [req-44060c58-97c5-4c23-976b-1c3419379bb5 req-7715148f-9be3-4697-b7b9-4a6bb0e9a981 service nova] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Received event network-vif-deleted-c1b660a5-8cfb-4650-9a8b-fa7bf0c491ab {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 658.565949] env[61243]: DEBUG nova.scheduler.client.report [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.601734] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Releasing lock "refresh_cache-4072d5a6-f69d-4fb9-837b-7855054e5206" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.602209] env[61243]: DEBUG nova.compute.manager [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 658.602418] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 658.603168] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-89ae1181-23d8-4647-be90-0bb53aee5444 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.612476] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b41cd278-b309-4ed4-a420-65dc6c58fdbf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.634835] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4072d5a6-f69d-4fb9-837b-7855054e5206 could not be found. [ 658.635067] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 658.635255] env[61243]: INFO nova.compute.manager [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Took 0.03 seconds to destroy the instance on the hypervisor. [ 658.635497] env[61243]: DEBUG oslo.service.loopingcall [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 658.635710] env[61243]: DEBUG nova.compute.manager [-] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.635805] env[61243]: DEBUG nova.network.neutron [-] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 658.653062] env[61243]: DEBUG nova.network.neutron [-] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.850278] env[61243]: ERROR nova.compute.manager [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0d346e84-b8e6-4678-99ad-bf03d7a94af8, please check neutron logs for more information. [ 658.850278] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 658.850278] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.850278] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 658.850278] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.850278] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 658.850278] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.850278] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 658.850278] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.850278] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 658.850278] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.850278] env[61243]: ERROR nova.compute.manager raise self.value [ 658.850278] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.850278] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 658.850278] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.850278] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 658.850859] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.850859] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 658.850859] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0d346e84-b8e6-4678-99ad-bf03d7a94af8, please check neutron logs for more information. [ 658.850859] env[61243]: ERROR nova.compute.manager [ 658.850859] env[61243]: Traceback (most recent call last): [ 658.850859] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 658.850859] env[61243]: listener.cb(fileno) [ 658.850859] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.850859] env[61243]: result = function(*args, **kwargs) [ 658.850859] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 658.850859] env[61243]: return func(*args, **kwargs) [ 658.850859] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.850859] env[61243]: raise e [ 658.850859] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.850859] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 658.850859] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.850859] env[61243]: created_port_ids = self._update_ports_for_instance( [ 658.850859] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.850859] env[61243]: with excutils.save_and_reraise_exception(): [ 658.850859] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.850859] env[61243]: self.force_reraise() [ 658.850859] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.850859] env[61243]: raise self.value [ 658.850859] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.850859] env[61243]: updated_port = self._update_port( [ 658.850859] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.850859] env[61243]: _ensure_no_port_binding_failure(port) [ 658.850859] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.850859] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 658.851704] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 0d346e84-b8e6-4678-99ad-bf03d7a94af8, please check neutron logs for more information. [ 658.851704] env[61243]: Removing descriptor: 16 [ 659.071822] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.926s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.072068] env[61243]: ERROR nova.compute.manager [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 49584d24-a0d0-45b8-af7c-bf7f39fb1964, please check neutron logs for more information. [ 659.072068] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Traceback (most recent call last): [ 659.072068] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 659.072068] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] self.driver.spawn(context, instance, image_meta, [ 659.072068] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 659.072068] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] self._vmops.spawn(context, instance, image_meta, injected_files, [ 659.072068] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 659.072068] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] vm_ref = self.build_virtual_machine(instance, [ 659.072068] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 659.072068] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] vif_infos = vmwarevif.get_vif_info(self._session, [ 659.072068] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 659.072417] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] for vif in network_info: [ 659.072417] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 659.072417] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] return self._sync_wrapper(fn, *args, **kwargs) [ 659.072417] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 659.072417] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] self.wait() [ 659.072417] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 659.072417] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] self[:] = self._gt.wait() [ 659.072417] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 659.072417] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] return self._exit_event.wait() [ 659.072417] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 659.072417] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] current.throw(*self._exc) [ 659.072417] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.072417] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] result = function(*args, **kwargs) [ 659.072882] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 659.072882] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] return func(*args, **kwargs) [ 659.072882] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.072882] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] raise e [ 659.072882] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.072882] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] nwinfo = self.network_api.allocate_for_instance( [ 659.072882] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.072882] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] created_port_ids = self._update_ports_for_instance( [ 659.072882] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.072882] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] with excutils.save_and_reraise_exception(): [ 659.072882] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.072882] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] self.force_reraise() [ 659.072882] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.073267] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] raise self.value [ 659.073267] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.073267] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] updated_port = self._update_port( [ 659.073267] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.073267] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] _ensure_no_port_binding_failure(port) [ 659.073267] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.073267] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] raise exception.PortBindingFailed(port_id=port['id']) [ 659.073267] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] nova.exception.PortBindingFailed: Binding failed for port 49584d24-a0d0-45b8-af7c-bf7f39fb1964, please check neutron logs for more information. [ 659.073267] env[61243]: ERROR nova.compute.manager [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] [ 659.073267] env[61243]: DEBUG nova.compute.utils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Binding failed for port 49584d24-a0d0-45b8-af7c-bf7f39fb1964, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 659.074052] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.547s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.075500] env[61243]: INFO nova.compute.claims [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 659.078769] env[61243]: DEBUG nova.compute.manager [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Build of instance b3d98e6f-4b45-4eaf-a79c-b683b32fe053 was re-scheduled: Binding failed for port 49584d24-a0d0-45b8-af7c-bf7f39fb1964, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 659.078960] env[61243]: DEBUG nova.compute.manager [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 659.079211] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Acquiring lock "refresh_cache-b3d98e6f-4b45-4eaf-a79c-b683b32fe053" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.079364] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Acquired lock "refresh_cache-b3d98e6f-4b45-4eaf-a79c-b683b32fe053" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.079527] env[61243]: DEBUG nova.network.neutron [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 659.154868] env[61243]: DEBUG nova.network.neutron [-] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.160228] env[61243]: DEBUG nova.compute.manager [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 659.186043] env[61243]: DEBUG nova.virt.hardware [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 659.186306] env[61243]: DEBUG nova.virt.hardware [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 659.186465] env[61243]: DEBUG nova.virt.hardware [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 659.186689] env[61243]: DEBUG nova.virt.hardware [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 659.186795] env[61243]: DEBUG nova.virt.hardware [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 659.186942] env[61243]: DEBUG nova.virt.hardware [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 659.187170] env[61243]: DEBUG nova.virt.hardware [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 659.187332] env[61243]: DEBUG nova.virt.hardware [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 659.187499] env[61243]: DEBUG nova.virt.hardware [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 659.187664] env[61243]: DEBUG nova.virt.hardware [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 659.187833] env[61243]: DEBUG nova.virt.hardware [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 659.188752] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34b5b7cd-23de-46ff-ab4c-9f4c385e9b8c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.197068] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f0b071-2e0d-4489-9b97-018c651f18ff {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.210673] env[61243]: ERROR nova.compute.manager [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0d346e84-b8e6-4678-99ad-bf03d7a94af8, please check neutron logs for more information. [ 659.210673] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Traceback (most recent call last): [ 659.210673] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 659.210673] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] yield resources [ 659.210673] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 659.210673] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] self.driver.spawn(context, instance, image_meta, [ 659.210673] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 659.210673] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] self._vmops.spawn(context, instance, image_meta, injected_files, [ 659.210673] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 659.210673] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] vm_ref = self.build_virtual_machine(instance, [ 659.210673] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 659.211087] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] vif_infos = vmwarevif.get_vif_info(self._session, [ 659.211087] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 659.211087] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] for vif in network_info: [ 659.211087] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 659.211087] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] return self._sync_wrapper(fn, *args, **kwargs) [ 659.211087] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 659.211087] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] self.wait() [ 659.211087] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 659.211087] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] self[:] = self._gt.wait() [ 659.211087] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 659.211087] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] return self._exit_event.wait() [ 659.211087] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 659.211087] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] current.throw(*self._exc) [ 659.211462] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.211462] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] result = function(*args, **kwargs) [ 659.211462] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 659.211462] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] return func(*args, **kwargs) [ 659.211462] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.211462] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] raise e [ 659.211462] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.211462] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] nwinfo = self.network_api.allocate_for_instance( [ 659.211462] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.211462] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] created_port_ids = self._update_ports_for_instance( [ 659.211462] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.211462] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] with excutils.save_and_reraise_exception(): [ 659.211462] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.211837] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] self.force_reraise() [ 659.211837] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.211837] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] raise self.value [ 659.211837] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.211837] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] updated_port = self._update_port( [ 659.211837] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.211837] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] _ensure_no_port_binding_failure(port) [ 659.211837] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.211837] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] raise exception.PortBindingFailed(port_id=port['id']) [ 659.211837] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] nova.exception.PortBindingFailed: Binding failed for port 0d346e84-b8e6-4678-99ad-bf03d7a94af8, please check neutron logs for more information. [ 659.211837] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] [ 659.211837] env[61243]: INFO nova.compute.manager [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Terminating instance [ 659.212968] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Acquiring lock "refresh_cache-d113d9ef-5698-4851-a51a-7c7c807e9135" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.213146] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Acquired lock "refresh_cache-d113d9ef-5698-4851-a51a-7c7c807e9135" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.213314] env[61243]: DEBUG nova.network.neutron [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 659.602398] env[61243]: DEBUG nova.network.neutron [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.656839] env[61243]: INFO nova.compute.manager [-] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Took 1.02 seconds to deallocate network for instance. [ 659.659706] env[61243]: DEBUG nova.compute.claims [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 659.659899] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.685237] env[61243]: DEBUG nova.network.neutron [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.731058] env[61243]: DEBUG nova.network.neutron [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.799502] env[61243]: DEBUG nova.network.neutron [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.187346] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Releasing lock "refresh_cache-b3d98e6f-4b45-4eaf-a79c-b683b32fe053" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.187626] env[61243]: DEBUG nova.compute.manager [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 660.187840] env[61243]: DEBUG nova.compute.manager [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 660.188052] env[61243]: DEBUG nova.network.neutron [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 660.202222] env[61243]: DEBUG nova.network.neutron [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.304017] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Releasing lock "refresh_cache-d113d9ef-5698-4851-a51a-7c7c807e9135" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.304017] env[61243]: DEBUG nova.compute.manager [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 660.304017] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 660.304417] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-34c5f25f-2317-4f8a-88b2-cbc8661b970f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.313089] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3cb8960-a331-48d0-84bf-9fe17e851034 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.337106] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d113d9ef-5698-4851-a51a-7c7c807e9135 could not be found. [ 660.337336] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 660.337518] env[61243]: INFO nova.compute.manager [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Took 0.03 seconds to destroy the instance on the hypervisor. [ 660.337757] env[61243]: DEBUG oslo.service.loopingcall [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 660.340085] env[61243]: DEBUG nova.compute.manager [-] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 660.340190] env[61243]: DEBUG nova.network.neutron [-] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 660.356775] env[61243]: DEBUG nova.network.neutron [-] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.420097] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15535fca-e99b-4a08-ae5f-25c45ee81d01 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.427313] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ec966b-b28c-423a-9a7f-ec3785e1189f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.458170] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4255ba43-92c2-469c-abbc-48198f75d086 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.462033] env[61243]: DEBUG nova.compute.manager [req-6f49791a-1558-4b48-9450-9c23f0290fcc req-11f4a41e-4c73-4ffa-ba40-ddf082941402 service nova] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Received event network-changed-0d346e84-b8e6-4678-99ad-bf03d7a94af8 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 660.462033] env[61243]: DEBUG nova.compute.manager [req-6f49791a-1558-4b48-9450-9c23f0290fcc req-11f4a41e-4c73-4ffa-ba40-ddf082941402 service nova] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Refreshing instance network info cache due to event network-changed-0d346e84-b8e6-4678-99ad-bf03d7a94af8. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 660.462264] env[61243]: DEBUG oslo_concurrency.lockutils [req-6f49791a-1558-4b48-9450-9c23f0290fcc req-11f4a41e-4c73-4ffa-ba40-ddf082941402 service nova] Acquiring lock "refresh_cache-d113d9ef-5698-4851-a51a-7c7c807e9135" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.462402] env[61243]: DEBUG oslo_concurrency.lockutils [req-6f49791a-1558-4b48-9450-9c23f0290fcc req-11f4a41e-4c73-4ffa-ba40-ddf082941402 service nova] Acquired lock "refresh_cache-d113d9ef-5698-4851-a51a-7c7c807e9135" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.462562] env[61243]: DEBUG nova.network.neutron [req-6f49791a-1558-4b48-9450-9c23f0290fcc req-11f4a41e-4c73-4ffa-ba40-ddf082941402 service nova] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Refreshing network info cache for port 0d346e84-b8e6-4678-99ad-bf03d7a94af8 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 660.470725] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8860f92a-8396-4208-87d2-50c48ec492f6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.486469] env[61243]: DEBUG nova.compute.provider_tree [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.705350] env[61243]: DEBUG nova.network.neutron [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.859957] env[61243]: DEBUG nova.network.neutron [-] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.980071] env[61243]: DEBUG nova.network.neutron [req-6f49791a-1558-4b48-9450-9c23f0290fcc req-11f4a41e-4c73-4ffa-ba40-ddf082941402 service nova] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.989318] env[61243]: DEBUG nova.scheduler.client.report [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.048050] env[61243]: DEBUG nova.network.neutron [req-6f49791a-1558-4b48-9450-9c23f0290fcc req-11f4a41e-4c73-4ffa-ba40-ddf082941402 service nova] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.207969] env[61243]: INFO nova.compute.manager [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] [instance: b3d98e6f-4b45-4eaf-a79c-b683b32fe053] Took 1.02 seconds to deallocate network for instance. [ 661.363561] env[61243]: INFO nova.compute.manager [-] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Took 1.02 seconds to deallocate network for instance. [ 661.366620] env[61243]: DEBUG nova.compute.claims [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 661.366620] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.494026] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.420s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.494262] env[61243]: DEBUG nova.compute.manager [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 661.496791] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.574s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.497252] env[61243]: DEBUG nova.objects.instance [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Lazy-loading 'resources' on Instance uuid b5163d89-ceb8-4c61-ae02-0ae3311b58a9 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 661.550598] env[61243]: DEBUG oslo_concurrency.lockutils [req-6f49791a-1558-4b48-9450-9c23f0290fcc req-11f4a41e-4c73-4ffa-ba40-ddf082941402 service nova] Releasing lock "refresh_cache-d113d9ef-5698-4851-a51a-7c7c807e9135" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.550825] env[61243]: DEBUG nova.compute.manager [req-6f49791a-1558-4b48-9450-9c23f0290fcc req-11f4a41e-4c73-4ffa-ba40-ddf082941402 service nova] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Received event network-vif-deleted-0d346e84-b8e6-4678-99ad-bf03d7a94af8 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 662.004413] env[61243]: DEBUG nova.compute.utils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 662.006233] env[61243]: DEBUG nova.compute.manager [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 662.007157] env[61243]: DEBUG nova.network.neutron [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 662.051123] env[61243]: DEBUG nova.policy [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '29275dd6899d4982bcacbdb35ae2de9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e697209eb99e4199891657e359f32690', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 662.236148] env[61243]: INFO nova.scheduler.client.report [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Deleted allocations for instance b3d98e6f-4b45-4eaf-a79c-b683b32fe053 [ 662.330514] env[61243]: DEBUG nova.network.neutron [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Successfully created port: 449658ee-44ec-497f-b8a3-54799b7ea061 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 662.357972] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d0bdbac-115d-4edb-961f-a5bb524e8e83 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.368724] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b74711c1-1cf5-4cac-a6b4-96d57a6714ef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.400210] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2dcdf98-18ce-4451-b5e6-3a81edb203d4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.408860] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-434942b1-d47a-4133-8600-132a7d42aa58 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.423027] env[61243]: DEBUG nova.compute.provider_tree [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.510186] env[61243]: DEBUG nova.compute.manager [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 662.744769] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab44eb55-360b-45d7-91d8-bb36adcf9636 tempest-ImagesOneServerNegativeTestJSON-1722038861 tempest-ImagesOneServerNegativeTestJSON-1722038861-project-member] Lock "b3d98e6f-4b45-4eaf-a79c-b683b32fe053" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.597s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.925853] env[61243]: DEBUG nova.scheduler.client.report [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.068167] env[61243]: DEBUG nova.compute.manager [req-a55003a0-6914-4011-9fd1-fd5bdc2484ea req-54682717-d24f-4f5d-a46b-3dac5013b7be service nova] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Received event network-changed-449658ee-44ec-497f-b8a3-54799b7ea061 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 663.068454] env[61243]: DEBUG nova.compute.manager [req-a55003a0-6914-4011-9fd1-fd5bdc2484ea req-54682717-d24f-4f5d-a46b-3dac5013b7be service nova] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Refreshing instance network info cache due to event network-changed-449658ee-44ec-497f-b8a3-54799b7ea061. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 663.068640] env[61243]: DEBUG oslo_concurrency.lockutils [req-a55003a0-6914-4011-9fd1-fd5bdc2484ea req-54682717-d24f-4f5d-a46b-3dac5013b7be service nova] Acquiring lock "refresh_cache-06264429-11eb-4d64-9f6b-6587cf048c41" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.068789] env[61243]: DEBUG oslo_concurrency.lockutils [req-a55003a0-6914-4011-9fd1-fd5bdc2484ea req-54682717-d24f-4f5d-a46b-3dac5013b7be service nova] Acquired lock "refresh_cache-06264429-11eb-4d64-9f6b-6587cf048c41" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.068951] env[61243]: DEBUG nova.network.neutron [req-a55003a0-6914-4011-9fd1-fd5bdc2484ea req-54682717-d24f-4f5d-a46b-3dac5013b7be service nova] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Refreshing network info cache for port 449658ee-44ec-497f-b8a3-54799b7ea061 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 663.180273] env[61243]: ERROR nova.compute.manager [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 449658ee-44ec-497f-b8a3-54799b7ea061, please check neutron logs for more information. [ 663.180273] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 663.180273] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.180273] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 663.180273] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.180273] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 663.180273] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.180273] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 663.180273] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.180273] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 663.180273] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.180273] env[61243]: ERROR nova.compute.manager raise self.value [ 663.180273] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.180273] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 663.180273] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.180273] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 663.180904] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.180904] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 663.180904] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 449658ee-44ec-497f-b8a3-54799b7ea061, please check neutron logs for more information. [ 663.180904] env[61243]: ERROR nova.compute.manager [ 663.180904] env[61243]: Traceback (most recent call last): [ 663.180904] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 663.180904] env[61243]: listener.cb(fileno) [ 663.180904] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.180904] env[61243]: result = function(*args, **kwargs) [ 663.180904] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 663.180904] env[61243]: return func(*args, **kwargs) [ 663.180904] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.180904] env[61243]: raise e [ 663.180904] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.180904] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 663.180904] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.180904] env[61243]: created_port_ids = self._update_ports_for_instance( [ 663.180904] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.180904] env[61243]: with excutils.save_and_reraise_exception(): [ 663.180904] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.180904] env[61243]: self.force_reraise() [ 663.180904] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.180904] env[61243]: raise self.value [ 663.180904] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.180904] env[61243]: updated_port = self._update_port( [ 663.180904] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.180904] env[61243]: _ensure_no_port_binding_failure(port) [ 663.180904] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.180904] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 663.182103] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 449658ee-44ec-497f-b8a3-54799b7ea061, please check neutron logs for more information. [ 663.182103] env[61243]: Removing descriptor: 16 [ 663.247041] env[61243]: DEBUG nova.compute.manager [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 663.432915] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.936s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.435260] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 20.316s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.435447] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.435610] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61243) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 663.435907] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.234s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.439540] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e505e7-4dc6-4d3f-bc8b-66eb65c0257e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.448371] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8144f3d4-2920-46ae-91e3-de0dcd366260 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.463801] env[61243]: INFO nova.scheduler.client.report [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Deleted allocations for instance b5163d89-ceb8-4c61-ae02-0ae3311b58a9 [ 663.465240] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e133369-71f9-4eee-9ee3-4afcc2a14628 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.476379] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc86dff4-28aa-4c5d-8234-f7a852470687 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.507024] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181519MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=61243) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 663.507233] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.520078] env[61243]: DEBUG nova.compute.manager [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 663.546156] env[61243]: DEBUG nova.virt.hardware [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 663.546408] env[61243]: DEBUG nova.virt.hardware [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 663.546568] env[61243]: DEBUG nova.virt.hardware [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 663.546754] env[61243]: DEBUG nova.virt.hardware [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 663.546899] env[61243]: DEBUG nova.virt.hardware [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 663.547064] env[61243]: DEBUG nova.virt.hardware [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 663.547383] env[61243]: DEBUG nova.virt.hardware [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 663.547452] env[61243]: DEBUG nova.virt.hardware [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 663.547589] env[61243]: DEBUG nova.virt.hardware [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 663.547752] env[61243]: DEBUG nova.virt.hardware [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 663.547992] env[61243]: DEBUG nova.virt.hardware [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 663.548878] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f90a03e-ae34-45da-9b32-0cf62da4cb80 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.557124] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f518cdf3-87dc-442b-aa7b-395fbfb6bada {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.571016] env[61243]: ERROR nova.compute.manager [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 449658ee-44ec-497f-b8a3-54799b7ea061, please check neutron logs for more information. [ 663.571016] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Traceback (most recent call last): [ 663.571016] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 663.571016] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] yield resources [ 663.571016] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.571016] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] self.driver.spawn(context, instance, image_meta, [ 663.571016] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 663.571016] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.571016] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.571016] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] vm_ref = self.build_virtual_machine(instance, [ 663.571016] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.571492] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.571492] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.571492] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] for vif in network_info: [ 663.571492] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.571492] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] return self._sync_wrapper(fn, *args, **kwargs) [ 663.571492] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.571492] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] self.wait() [ 663.571492] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.571492] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] self[:] = self._gt.wait() [ 663.571492] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.571492] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] return self._exit_event.wait() [ 663.571492] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 663.571492] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] current.throw(*self._exc) [ 663.572214] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.572214] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] result = function(*args, **kwargs) [ 663.572214] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 663.572214] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] return func(*args, **kwargs) [ 663.572214] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.572214] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] raise e [ 663.572214] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.572214] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] nwinfo = self.network_api.allocate_for_instance( [ 663.572214] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.572214] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] created_port_ids = self._update_ports_for_instance( [ 663.572214] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.572214] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] with excutils.save_and_reraise_exception(): [ 663.572214] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.572783] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] self.force_reraise() [ 663.572783] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.572783] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] raise self.value [ 663.572783] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.572783] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] updated_port = self._update_port( [ 663.572783] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.572783] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] _ensure_no_port_binding_failure(port) [ 663.572783] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.572783] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] raise exception.PortBindingFailed(port_id=port['id']) [ 663.572783] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] nova.exception.PortBindingFailed: Binding failed for port 449658ee-44ec-497f-b8a3-54799b7ea061, please check neutron logs for more information. [ 663.572783] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] [ 663.572783] env[61243]: INFO nova.compute.manager [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Terminating instance [ 663.574691] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Acquiring lock "refresh_cache-06264429-11eb-4d64-9f6b-6587cf048c41" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.588789] env[61243]: DEBUG nova.network.neutron [req-a55003a0-6914-4011-9fd1-fd5bdc2484ea req-54682717-d24f-4f5d-a46b-3dac5013b7be service nova] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.664072] env[61243]: DEBUG nova.network.neutron [req-a55003a0-6914-4011-9fd1-fd5bdc2484ea req-54682717-d24f-4f5d-a46b-3dac5013b7be service nova] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.765099] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.977067] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8912d74d-221a-48f3-af67-6dcab2da3b66 tempest-ServerDiagnosticsV248Test-1911283959 tempest-ServerDiagnosticsV248Test-1911283959-project-member] Lock "b5163d89-ceb8-4c61-ae02-0ae3311b58a9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.761s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.168412] env[61243]: DEBUG oslo_concurrency.lockutils [req-a55003a0-6914-4011-9fd1-fd5bdc2484ea req-54682717-d24f-4f5d-a46b-3dac5013b7be service nova] Releasing lock "refresh_cache-06264429-11eb-4d64-9f6b-6587cf048c41" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.169039] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Acquired lock "refresh_cache-06264429-11eb-4d64-9f6b-6587cf048c41" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.169241] env[61243]: DEBUG nova.network.neutron [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 664.328162] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3eb398-5f0e-4f48-98e4-27850e595f22 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.335872] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc28398-8a71-4441-b1b3-869a3d6ec454 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.366134] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa8c45c2-32be-499f-8c7e-adf1ae6aef7e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.377536] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fba32e7-37f3-44b3-9185-0d678c30ffd6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.392381] env[61243]: DEBUG nova.compute.provider_tree [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.687937] env[61243]: DEBUG nova.network.neutron [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.830963] env[61243]: DEBUG nova.network.neutron [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.896699] env[61243]: DEBUG nova.scheduler.client.report [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 665.132256] env[61243]: DEBUG nova.compute.manager [req-92d396f9-c6e3-46fb-8883-57361e3d066a req-567eeb05-27d9-4095-ac30-1bd754d3dba6 service nova] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Received event network-vif-deleted-449658ee-44ec-497f-b8a3-54799b7ea061 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 665.335047] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Releasing lock "refresh_cache-06264429-11eb-4d64-9f6b-6587cf048c41" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.335047] env[61243]: DEBUG nova.compute.manager [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 665.335047] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 665.335047] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b911cd35-24eb-4da2-8762-04fe603884fb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.345043] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f2fe3d-aa48-46a1-9c37-d1d416f80c9a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.370217] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 06264429-11eb-4d64-9f6b-6587cf048c41 could not be found. [ 665.370217] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 665.370217] env[61243]: INFO nova.compute.manager [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Took 0.03 seconds to destroy the instance on the hypervisor. [ 665.370217] env[61243]: DEBUG oslo.service.loopingcall [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 665.370217] env[61243]: DEBUG nova.compute.manager [-] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 665.370217] env[61243]: DEBUG nova.network.neutron [-] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 665.388158] env[61243]: DEBUG nova.network.neutron [-] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.402545] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.966s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.403161] env[61243]: ERROR nova.compute.manager [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f3056083-ab3a-48ad-ae6c-37e5fa1b3d60, please check neutron logs for more information. [ 665.403161] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Traceback (most recent call last): [ 665.403161] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.403161] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] self.driver.spawn(context, instance, image_meta, [ 665.403161] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 665.403161] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.403161] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.403161] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] vm_ref = self.build_virtual_machine(instance, [ 665.403161] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.403161] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.403161] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.403492] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] for vif in network_info: [ 665.403492] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.403492] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] return self._sync_wrapper(fn, *args, **kwargs) [ 665.403492] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.403492] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] self.wait() [ 665.403492] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.403492] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] self[:] = self._gt.wait() [ 665.403492] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.403492] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] return self._exit_event.wait() [ 665.403492] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 665.403492] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] current.throw(*self._exc) [ 665.403492] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.403492] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] result = function(*args, **kwargs) [ 665.403911] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 665.403911] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] return func(*args, **kwargs) [ 665.403911] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.403911] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] raise e [ 665.403911] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.403911] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] nwinfo = self.network_api.allocate_for_instance( [ 665.403911] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.403911] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] created_port_ids = self._update_ports_for_instance( [ 665.403911] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.403911] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] with excutils.save_and_reraise_exception(): [ 665.403911] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.403911] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] self.force_reraise() [ 665.403911] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.404229] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] raise self.value [ 665.404229] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.404229] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] updated_port = self._update_port( [ 665.404229] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.404229] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] _ensure_no_port_binding_failure(port) [ 665.404229] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.404229] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] raise exception.PortBindingFailed(port_id=port['id']) [ 665.404229] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] nova.exception.PortBindingFailed: Binding failed for port f3056083-ab3a-48ad-ae6c-37e5fa1b3d60, please check neutron logs for more information. [ 665.404229] env[61243]: ERROR nova.compute.manager [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] [ 665.404229] env[61243]: DEBUG nova.compute.utils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Binding failed for port f3056083-ab3a-48ad-ae6c-37e5fa1b3d60, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 665.405129] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.394s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.406629] env[61243]: INFO nova.compute.claims [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 665.409424] env[61243]: DEBUG nova.compute.manager [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Build of instance 0b223791-7e95-43fe-bef0-bac6222884ea was re-scheduled: Binding failed for port f3056083-ab3a-48ad-ae6c-37e5fa1b3d60, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 665.410201] env[61243]: DEBUG nova.compute.manager [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 665.410201] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Acquiring lock "refresh_cache-0b223791-7e95-43fe-bef0-bac6222884ea" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.410307] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Acquired lock "refresh_cache-0b223791-7e95-43fe-bef0-bac6222884ea" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.410388] env[61243]: DEBUG nova.network.neutron [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 665.890664] env[61243]: DEBUG nova.network.neutron [-] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.934750] env[61243]: DEBUG nova.network.neutron [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.065279] env[61243]: DEBUG nova.network.neutron [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.393804] env[61243]: INFO nova.compute.manager [-] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Took 1.02 seconds to deallocate network for instance. [ 666.400377] env[61243]: DEBUG nova.compute.claims [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 666.400580] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.568339] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Releasing lock "refresh_cache-0b223791-7e95-43fe-bef0-bac6222884ea" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.568576] env[61243]: DEBUG nova.compute.manager [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 666.568764] env[61243]: DEBUG nova.compute.manager [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 666.568920] env[61243]: DEBUG nova.network.neutron [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 666.654172] env[61243]: DEBUG nova.network.neutron [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.847533] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a24bf783-eb03-48b3-bf9a-40fbaad71362 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.860372] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dc0ee12-3a0d-44a4-97c1-002fb15d01ab {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.898465] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d1aef1-27e0-4e3f-94ba-6f540624b92d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.905972] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb8b891f-d041-4d6e-a9aa-4249a7fb0852 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.920574] env[61243]: DEBUG nova.compute.provider_tree [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.158250] env[61243]: DEBUG nova.network.neutron [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.423736] env[61243]: DEBUG nova.scheduler.client.report [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 667.666287] env[61243]: INFO nova.compute.manager [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] [instance: 0b223791-7e95-43fe-bef0-bac6222884ea] Took 1.09 seconds to deallocate network for instance. [ 667.928978] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.524s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.929923] env[61243]: DEBUG nova.compute.manager [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 667.932591] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.678s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.440016] env[61243]: DEBUG nova.compute.utils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 668.444075] env[61243]: DEBUG nova.compute.manager [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 668.446484] env[61243]: DEBUG nova.network.neutron [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 668.493621] env[61243]: DEBUG nova.policy [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cc77f359ba44dc0bc8e053d5a4b63ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3248cca8a8aa4cad84b8bd5dcf7ed0c3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 668.694957] env[61243]: INFO nova.scheduler.client.report [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Deleted allocations for instance 0b223791-7e95-43fe-bef0-bac6222884ea [ 668.783427] env[61243]: DEBUG nova.network.neutron [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Successfully created port: 6fb89631-00db-453d-999a-a666fbd00063 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 668.875803] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8040ff79-76c4-42bc-b45a-895eba560508 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.885522] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f247451b-fa91-4c58-befe-b42f58f00bc6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.921119] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6eb3036-3443-4b38-9e79-3f31b49f85d4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.930338] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd026ad0-0ed8-4742-8e47-c58985b25232 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.945903] env[61243]: DEBUG nova.compute.manager [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 668.949136] env[61243]: DEBUG nova.compute.provider_tree [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.205356] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3ce591-15ff-4adb-bfd3-794de3938e8c tempest-ImagesOneServerTestJSON-1526135327 tempest-ImagesOneServerTestJSON-1526135327-project-member] Lock "0b223791-7e95-43fe-bef0-bac6222884ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.550s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.455534] env[61243]: DEBUG nova.scheduler.client.report [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 669.623017] env[61243]: DEBUG nova.compute.manager [req-ef252613-bdc3-4fab-afc6-610f30f35fea req-37421503-a384-4720-84a8-e182fbdc93e2 service nova] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Received event network-changed-6fb89631-00db-453d-999a-a666fbd00063 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 669.623017] env[61243]: DEBUG nova.compute.manager [req-ef252613-bdc3-4fab-afc6-610f30f35fea req-37421503-a384-4720-84a8-e182fbdc93e2 service nova] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Refreshing instance network info cache due to event network-changed-6fb89631-00db-453d-999a-a666fbd00063. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 669.623017] env[61243]: DEBUG oslo_concurrency.lockutils [req-ef252613-bdc3-4fab-afc6-610f30f35fea req-37421503-a384-4720-84a8-e182fbdc93e2 service nova] Acquiring lock "refresh_cache-d5131a9f-f4bd-40c8-98d1-043651537a16" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.623017] env[61243]: DEBUG oslo_concurrency.lockutils [req-ef252613-bdc3-4fab-afc6-610f30f35fea req-37421503-a384-4720-84a8-e182fbdc93e2 service nova] Acquired lock "refresh_cache-d5131a9f-f4bd-40c8-98d1-043651537a16" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.623017] env[61243]: DEBUG nova.network.neutron [req-ef252613-bdc3-4fab-afc6-610f30f35fea req-37421503-a384-4720-84a8-e182fbdc93e2 service nova] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Refreshing network info cache for port 6fb89631-00db-453d-999a-a666fbd00063 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 669.683932] env[61243]: ERROR nova.compute.manager [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6fb89631-00db-453d-999a-a666fbd00063, please check neutron logs for more information. [ 669.683932] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 669.683932] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.683932] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 669.683932] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.683932] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 669.683932] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.683932] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 669.683932] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.683932] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 669.683932] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.683932] env[61243]: ERROR nova.compute.manager raise self.value [ 669.683932] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.683932] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 669.683932] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.683932] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 669.684535] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.684535] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 669.684535] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6fb89631-00db-453d-999a-a666fbd00063, please check neutron logs for more information. [ 669.684535] env[61243]: ERROR nova.compute.manager [ 669.685000] env[61243]: Traceback (most recent call last): [ 669.685000] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 669.685000] env[61243]: listener.cb(fileno) [ 669.685000] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.685000] env[61243]: result = function(*args, **kwargs) [ 669.685000] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 669.685000] env[61243]: return func(*args, **kwargs) [ 669.685000] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.685000] env[61243]: raise e [ 669.685000] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.685000] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 669.685000] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.685000] env[61243]: created_port_ids = self._update_ports_for_instance( [ 669.685000] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.685000] env[61243]: with excutils.save_and_reraise_exception(): [ 669.685000] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.685000] env[61243]: self.force_reraise() [ 669.685000] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.685000] env[61243]: raise self.value [ 669.685000] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.685000] env[61243]: updated_port = self._update_port( [ 669.685000] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.685000] env[61243]: _ensure_no_port_binding_failure(port) [ 669.685000] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.685000] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 669.685000] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 6fb89631-00db-453d-999a-a666fbd00063, please check neutron logs for more information. [ 669.685000] env[61243]: Removing descriptor: 16 [ 669.709698] env[61243]: DEBUG nova.compute.manager [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 669.960222] env[61243]: DEBUG nova.compute.manager [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 669.962939] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.030s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.963561] env[61243]: ERROR nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 66e28bb3-150d-4a9f-8b17-b7c1911a40cd, please check neutron logs for more information. [ 669.963561] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Traceback (most recent call last): [ 669.963561] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 669.963561] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] self.driver.spawn(context, instance, image_meta, [ 669.963561] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 669.963561] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.963561] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.963561] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] vm_ref = self.build_virtual_machine(instance, [ 669.963561] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.963561] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.963561] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.963898] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] for vif in network_info: [ 669.963898] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.963898] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] return self._sync_wrapper(fn, *args, **kwargs) [ 669.963898] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.963898] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] self.wait() [ 669.963898] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.963898] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] self[:] = self._gt.wait() [ 669.963898] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.963898] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] return self._exit_event.wait() [ 669.963898] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 669.963898] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] current.throw(*self._exc) [ 669.963898] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.963898] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] result = function(*args, **kwargs) [ 669.964227] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 669.964227] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] return func(*args, **kwargs) [ 669.964227] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.964227] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] raise e [ 669.964227] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.964227] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] nwinfo = self.network_api.allocate_for_instance( [ 669.964227] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.964227] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] created_port_ids = self._update_ports_for_instance( [ 669.964227] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.964227] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] with excutils.save_and_reraise_exception(): [ 669.964227] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.964227] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] self.force_reraise() [ 669.964227] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.964589] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] raise self.value [ 669.964589] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.964589] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] updated_port = self._update_port( [ 669.964589] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.964589] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] _ensure_no_port_binding_failure(port) [ 669.964589] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.964589] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] raise exception.PortBindingFailed(port_id=port['id']) [ 669.964589] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] nova.exception.PortBindingFailed: Binding failed for port 66e28bb3-150d-4a9f-8b17-b7c1911a40cd, please check neutron logs for more information. [ 669.964589] env[61243]: ERROR nova.compute.manager [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] [ 669.964589] env[61243]: DEBUG nova.compute.utils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Binding failed for port 66e28bb3-150d-4a9f-8b17-b7c1911a40cd, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 669.965515] env[61243]: DEBUG oslo_concurrency.lockutils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.532s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.972353] env[61243]: DEBUG nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Build of instance 0625ac10-55b4-488a-aba5-bdaa394045f4 was re-scheduled: Binding failed for port 66e28bb3-150d-4a9f-8b17-b7c1911a40cd, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 669.972832] env[61243]: DEBUG nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 669.973089] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Acquiring lock "refresh_cache-0625ac10-55b4-488a-aba5-bdaa394045f4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.973253] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Acquired lock "refresh_cache-0625ac10-55b4-488a-aba5-bdaa394045f4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.973438] env[61243]: DEBUG nova.network.neutron [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 669.997337] env[61243]: DEBUG nova.virt.hardware [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 669.997607] env[61243]: DEBUG nova.virt.hardware [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 669.997746] env[61243]: DEBUG nova.virt.hardware [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 669.997926] env[61243]: DEBUG nova.virt.hardware [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 669.998086] env[61243]: DEBUG nova.virt.hardware [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 669.998238] env[61243]: DEBUG nova.virt.hardware [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 669.998628] env[61243]: DEBUG nova.virt.hardware [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 669.998628] env[61243]: DEBUG nova.virt.hardware [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 669.998760] env[61243]: DEBUG nova.virt.hardware [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 669.998915] env[61243]: DEBUG nova.virt.hardware [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 669.999201] env[61243]: DEBUG nova.virt.hardware [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 670.000371] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f3589ae-b0c6-4bad-9504-5d1078a06de1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.010074] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478031b1-02c0-46cf-8e6b-fc6c2b17f76a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.026994] env[61243]: ERROR nova.compute.manager [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6fb89631-00db-453d-999a-a666fbd00063, please check neutron logs for more information. [ 670.026994] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Traceback (most recent call last): [ 670.026994] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 670.026994] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] yield resources [ 670.026994] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.026994] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] self.driver.spawn(context, instance, image_meta, [ 670.026994] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 670.026994] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.026994] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.026994] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] vm_ref = self.build_virtual_machine(instance, [ 670.026994] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.027477] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.027477] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.027477] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] for vif in network_info: [ 670.027477] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.027477] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] return self._sync_wrapper(fn, *args, **kwargs) [ 670.027477] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.027477] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] self.wait() [ 670.027477] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.027477] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] self[:] = self._gt.wait() [ 670.027477] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.027477] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] return self._exit_event.wait() [ 670.027477] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 670.027477] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] current.throw(*self._exc) [ 670.027834] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.027834] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] result = function(*args, **kwargs) [ 670.027834] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 670.027834] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] return func(*args, **kwargs) [ 670.027834] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.027834] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] raise e [ 670.027834] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.027834] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] nwinfo = self.network_api.allocate_for_instance( [ 670.027834] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.027834] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] created_port_ids = self._update_ports_for_instance( [ 670.027834] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.027834] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] with excutils.save_and_reraise_exception(): [ 670.027834] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.028232] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] self.force_reraise() [ 670.028232] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.028232] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] raise self.value [ 670.028232] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.028232] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] updated_port = self._update_port( [ 670.028232] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.028232] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] _ensure_no_port_binding_failure(port) [ 670.028232] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.028232] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] raise exception.PortBindingFailed(port_id=port['id']) [ 670.028232] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] nova.exception.PortBindingFailed: Binding failed for port 6fb89631-00db-453d-999a-a666fbd00063, please check neutron logs for more information. [ 670.028232] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] [ 670.028232] env[61243]: INFO nova.compute.manager [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Terminating instance [ 670.029446] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "refresh_cache-d5131a9f-f4bd-40c8-98d1-043651537a16" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.139840] env[61243]: DEBUG nova.network.neutron [req-ef252613-bdc3-4fab-afc6-610f30f35fea req-37421503-a384-4720-84a8-e182fbdc93e2 service nova] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.236640] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.236640] env[61243]: DEBUG nova.network.neutron [req-ef252613-bdc3-4fab-afc6-610f30f35fea req-37421503-a384-4720-84a8-e182fbdc93e2 service nova] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.506985] env[61243]: DEBUG nova.network.neutron [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.632183] env[61243]: DEBUG nova.network.neutron [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.742822] env[61243]: DEBUG oslo_concurrency.lockutils [req-ef252613-bdc3-4fab-afc6-610f30f35fea req-37421503-a384-4720-84a8-e182fbdc93e2 service nova] Releasing lock "refresh_cache-d5131a9f-f4bd-40c8-98d1-043651537a16" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.743482] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired lock "refresh_cache-d5131a9f-f4bd-40c8-98d1-043651537a16" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.746469] env[61243]: DEBUG nova.network.neutron [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 670.918351] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c9ad6a1-cc7c-4398-a2be-c434d00c275b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.926807] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee0f6b6-db72-4805-97ca-4808a136131d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.965159] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f1bf63-cb5a-4d5b-b818-41b386e3d2da {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.974478] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d09f32-86b4-4348-b9f8-725a7ecb7262 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.988992] env[61243]: DEBUG nova.compute.provider_tree [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.140395] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Releasing lock "refresh_cache-0625ac10-55b4-488a-aba5-bdaa394045f4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.140807] env[61243]: DEBUG nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 671.141482] env[61243]: DEBUG nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.143248] env[61243]: DEBUG nova.network.neutron [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 671.165156] env[61243]: DEBUG nova.network.neutron [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.265350] env[61243]: DEBUG nova.network.neutron [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.352418] env[61243]: DEBUG nova.network.neutron [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.496121] env[61243]: DEBUG nova.scheduler.client.report [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 671.670080] env[61243]: DEBUG nova.network.neutron [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.855818] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Releasing lock "refresh_cache-d5131a9f-f4bd-40c8-98d1-043651537a16" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.857409] env[61243]: DEBUG nova.compute.manager [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 671.857409] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 671.857866] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd845fc0-558c-468b-81ca-52286615f211 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.869726] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71c68be-0adf-4cd2-813d-85f967cca02e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.889017] env[61243]: DEBUG nova.compute.manager [req-18cd13dc-caa7-46c0-8f0f-b6054cbb4ea3 req-02a1328e-2e0d-4ada-83a3-3a30a0e2ab73 service nova] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Received event network-vif-deleted-6fb89631-00db-453d-999a-a666fbd00063 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 671.905110] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d5131a9f-f4bd-40c8-98d1-043651537a16 could not be found. [ 671.905110] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 671.905110] env[61243]: INFO nova.compute.manager [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Took 0.05 seconds to destroy the instance on the hypervisor. [ 671.905597] env[61243]: DEBUG oslo.service.loopingcall [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 671.905878] env[61243]: DEBUG nova.compute.manager [-] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.905942] env[61243]: DEBUG nova.network.neutron [-] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 671.925468] env[61243]: DEBUG nova.network.neutron [-] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.003209] env[61243]: DEBUG oslo_concurrency.lockutils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.037s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.004529] env[61243]: ERROR nova.compute.manager [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e148ba07-8dcb-4345-93fa-15a43fbbbe3f, please check neutron logs for more information. [ 672.004529] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Traceback (most recent call last): [ 672.004529] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 672.004529] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] self.driver.spawn(context, instance, image_meta, [ 672.004529] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 672.004529] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.004529] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.004529] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] vm_ref = self.build_virtual_machine(instance, [ 672.004529] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.004529] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.004529] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.004811] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] for vif in network_info: [ 672.004811] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 672.004811] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] return self._sync_wrapper(fn, *args, **kwargs) [ 672.004811] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 672.004811] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] self.wait() [ 672.004811] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 672.004811] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] self[:] = self._gt.wait() [ 672.004811] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.004811] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] return self._exit_event.wait() [ 672.004811] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 672.004811] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] current.throw(*self._exc) [ 672.004811] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.004811] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] result = function(*args, **kwargs) [ 672.005147] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 672.005147] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] return func(*args, **kwargs) [ 672.005147] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.005147] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] raise e [ 672.005147] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.005147] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] nwinfo = self.network_api.allocate_for_instance( [ 672.005147] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.005147] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] created_port_ids = self._update_ports_for_instance( [ 672.005147] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.005147] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] with excutils.save_and_reraise_exception(): [ 672.005147] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.005147] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] self.force_reraise() [ 672.005147] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.005453] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] raise self.value [ 672.005453] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.005453] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] updated_port = self._update_port( [ 672.005453] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.005453] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] _ensure_no_port_binding_failure(port) [ 672.005453] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.005453] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] raise exception.PortBindingFailed(port_id=port['id']) [ 672.005453] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] nova.exception.PortBindingFailed: Binding failed for port e148ba07-8dcb-4345-93fa-15a43fbbbe3f, please check neutron logs for more information. [ 672.005453] env[61243]: ERROR nova.compute.manager [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] [ 672.005453] env[61243]: DEBUG nova.compute.utils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Binding failed for port e148ba07-8dcb-4345-93fa-15a43fbbbe3f, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 672.007285] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.102s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.011391] env[61243]: DEBUG nova.compute.manager [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Build of instance 64efd676-9166-4ffa-a437-6fa6c466ace0 was re-scheduled: Binding failed for port e148ba07-8dcb-4345-93fa-15a43fbbbe3f, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 672.011855] env[61243]: DEBUG nova.compute.manager [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 672.012154] env[61243]: DEBUG oslo_concurrency.lockutils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Acquiring lock "refresh_cache-64efd676-9166-4ffa-a437-6fa6c466ace0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.012354] env[61243]: DEBUG oslo_concurrency.lockutils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Acquired lock "refresh_cache-64efd676-9166-4ffa-a437-6fa6c466ace0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.012562] env[61243]: DEBUG nova.network.neutron [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 672.173134] env[61243]: INFO nova.compute.manager [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] [instance: 0625ac10-55b4-488a-aba5-bdaa394045f4] Took 1.03 seconds to deallocate network for instance. [ 672.427180] env[61243]: DEBUG nova.network.neutron [-] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.436648] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Acquiring lock "5314f922-6603-40dd-8760-9c1b77d6bd0a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.436648] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Lock "5314f922-6603-40dd-8760-9c1b77d6bd0a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.540406] env[61243]: DEBUG nova.network.neutron [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.623373] env[61243]: DEBUG nova.network.neutron [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.901351] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49112e64-1194-4436-82ae-6ffc16a1eaa0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.909996] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b865ec-3b1c-4e7f-a949-f5f34616bb86 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.943230] env[61243]: INFO nova.compute.manager [-] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Took 1.04 seconds to deallocate network for instance. [ 672.946562] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ac089b9-907b-4466-9d51-e8b7e0ded3f1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.948731] env[61243]: DEBUG nova.compute.claims [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 672.948984] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.957455] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-504040b7-92c4-4041-98df-678ff968c20f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.971890] env[61243]: DEBUG nova.compute.provider_tree [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.126389] env[61243]: DEBUG oslo_concurrency.lockutils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Releasing lock "refresh_cache-64efd676-9166-4ffa-a437-6fa6c466ace0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.126643] env[61243]: DEBUG nova.compute.manager [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 673.126829] env[61243]: DEBUG nova.compute.manager [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 673.126995] env[61243]: DEBUG nova.network.neutron [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 673.143761] env[61243]: DEBUG nova.network.neutron [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.227124] env[61243]: INFO nova.scheduler.client.report [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Deleted allocations for instance 0625ac10-55b4-488a-aba5-bdaa394045f4 [ 673.477864] env[61243]: DEBUG nova.scheduler.client.report [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 673.646648] env[61243]: DEBUG nova.network.neutron [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.736256] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10f2c2d5-3c24-4cbc-b4fc-37dd7e48ca5b tempest-ServerActionsV293TestJSON-1790371455 tempest-ServerActionsV293TestJSON-1790371455-project-member] Lock "0625ac10-55b4-488a-aba5-bdaa394045f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.905s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.980694] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.973s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.981556] env[61243]: ERROR nova.compute.manager [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a073df5c-76be-4a85-b48a-b7a65a1d7482, please check neutron logs for more information. [ 673.981556] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Traceback (most recent call last): [ 673.981556] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 673.981556] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] self.driver.spawn(context, instance, image_meta, [ 673.981556] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 673.981556] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 673.981556] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 673.981556] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] vm_ref = self.build_virtual_machine(instance, [ 673.981556] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 673.981556] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] vif_infos = vmwarevif.get_vif_info(self._session, [ 673.981556] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 673.982096] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] for vif in network_info: [ 673.982096] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 673.982096] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] return self._sync_wrapper(fn, *args, **kwargs) [ 673.982096] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 673.982096] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] self.wait() [ 673.982096] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 673.982096] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] self[:] = self._gt.wait() [ 673.982096] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 673.982096] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] return self._exit_event.wait() [ 673.982096] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 673.982096] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] current.throw(*self._exc) [ 673.982096] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 673.982096] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] result = function(*args, **kwargs) [ 673.982417] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 673.982417] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] return func(*args, **kwargs) [ 673.982417] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 673.982417] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] raise e [ 673.982417] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 673.982417] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] nwinfo = self.network_api.allocate_for_instance( [ 673.982417] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 673.982417] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] created_port_ids = self._update_ports_for_instance( [ 673.982417] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 673.982417] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] with excutils.save_and_reraise_exception(): [ 673.982417] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 673.982417] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] self.force_reraise() [ 673.982417] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 673.982733] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] raise self.value [ 673.982733] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 673.982733] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] updated_port = self._update_port( [ 673.982733] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 673.982733] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] _ensure_no_port_binding_failure(port) [ 673.982733] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 673.982733] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] raise exception.PortBindingFailed(port_id=port['id']) [ 673.982733] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] nova.exception.PortBindingFailed: Binding failed for port a073df5c-76be-4a85-b48a-b7a65a1d7482, please check neutron logs for more information. [ 673.982733] env[61243]: ERROR nova.compute.manager [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] [ 673.982733] env[61243]: DEBUG nova.compute.utils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Binding failed for port a073df5c-76be-4a85-b48a-b7a65a1d7482, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 673.983834] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.621s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.989211] env[61243]: DEBUG nova.compute.manager [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Build of instance 8efd391b-7b76-4167-bd18-48566271ccc8 was re-scheduled: Binding failed for port a073df5c-76be-4a85-b48a-b7a65a1d7482, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 673.989211] env[61243]: DEBUG nova.compute.manager [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 673.989211] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Acquiring lock "refresh_cache-8efd391b-7b76-4167-bd18-48566271ccc8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.989211] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Acquired lock "refresh_cache-8efd391b-7b76-4167-bd18-48566271ccc8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.989405] env[61243]: DEBUG nova.network.neutron [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 674.149247] env[61243]: INFO nova.compute.manager [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] [instance: 64efd676-9166-4ffa-a437-6fa6c466ace0] Took 1.02 seconds to deallocate network for instance. [ 674.238760] env[61243]: DEBUG nova.compute.manager [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 674.516102] env[61243]: DEBUG nova.network.neutron [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.673736] env[61243]: DEBUG nova.network.neutron [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.763832] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.941805] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be231e6-b280-4983-997e-254cf1964d0d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.952567] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f785c0-e0e0-4fa0-80f3-977d144e5611 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.000465] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44987fde-bc56-4519-8497-88804caeed6a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.008503] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53167935-77e5-40fa-8f70-a753fd11c8b1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.022746] env[61243]: DEBUG nova.compute.provider_tree [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.180519] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Releasing lock "refresh_cache-8efd391b-7b76-4167-bd18-48566271ccc8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.180519] env[61243]: DEBUG nova.compute.manager [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 675.180519] env[61243]: DEBUG nova.compute.manager [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 675.180519] env[61243]: DEBUG nova.network.neutron [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 675.204977] env[61243]: DEBUG nova.network.neutron [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.212340] env[61243]: INFO nova.scheduler.client.report [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Deleted allocations for instance 64efd676-9166-4ffa-a437-6fa6c466ace0 [ 675.526248] env[61243]: DEBUG nova.scheduler.client.report [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.710513] env[61243]: DEBUG nova.network.neutron [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.724487] env[61243]: DEBUG oslo_concurrency.lockutils [None req-813d5a06-eaca-4ad3-ae8f-cf09f20f9a3a tempest-ServerTagsTestJSON-1508722185 tempest-ServerTagsTestJSON-1508722185-project-member] Lock "64efd676-9166-4ffa-a437-6fa6c466ace0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.609s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.031187] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.047s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.033382] env[61243]: ERROR nova.compute.manager [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6eb50563-62d3-4e2a-8adf-9d0d5417e4f7, please check neutron logs for more information. [ 676.033382] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Traceback (most recent call last): [ 676.033382] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 676.033382] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] self.driver.spawn(context, instance, image_meta, [ 676.033382] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 676.033382] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 676.033382] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 676.033382] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] vm_ref = self.build_virtual_machine(instance, [ 676.033382] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 676.033382] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] vif_infos = vmwarevif.get_vif_info(self._session, [ 676.033382] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 676.034272] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] for vif in network_info: [ 676.034272] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 676.034272] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] return self._sync_wrapper(fn, *args, **kwargs) [ 676.034272] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 676.034272] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] self.wait() [ 676.034272] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 676.034272] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] self[:] = self._gt.wait() [ 676.034272] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 676.034272] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] return self._exit_event.wait() [ 676.034272] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 676.034272] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] current.throw(*self._exc) [ 676.034272] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.034272] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] result = function(*args, **kwargs) [ 676.035081] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 676.035081] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] return func(*args, **kwargs) [ 676.035081] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.035081] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] raise e [ 676.035081] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.035081] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] nwinfo = self.network_api.allocate_for_instance( [ 676.035081] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 676.035081] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] created_port_ids = self._update_ports_for_instance( [ 676.035081] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 676.035081] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] with excutils.save_and_reraise_exception(): [ 676.035081] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.035081] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] self.force_reraise() [ 676.035081] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.036110] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] raise self.value [ 676.036110] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 676.036110] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] updated_port = self._update_port( [ 676.036110] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.036110] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] _ensure_no_port_binding_failure(port) [ 676.036110] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.036110] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] raise exception.PortBindingFailed(port_id=port['id']) [ 676.036110] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] nova.exception.PortBindingFailed: Binding failed for port 6eb50563-62d3-4e2a-8adf-9d0d5417e4f7, please check neutron logs for more information. [ 676.036110] env[61243]: ERROR nova.compute.manager [instance: 84529438-8560-4323-a2f8-86cb917d18e6] [ 676.036110] env[61243]: DEBUG nova.compute.utils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Binding failed for port 6eb50563-62d3-4e2a-8adf-9d0d5417e4f7, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 676.037467] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.377s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.040803] env[61243]: DEBUG nova.compute.manager [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Build of instance 84529438-8560-4323-a2f8-86cb917d18e6 was re-scheduled: Binding failed for port 6eb50563-62d3-4e2a-8adf-9d0d5417e4f7, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 676.042032] env[61243]: DEBUG nova.compute.manager [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 676.042032] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Acquiring lock "refresh_cache-84529438-8560-4323-a2f8-86cb917d18e6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.042032] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Acquired lock "refresh_cache-84529438-8560-4323-a2f8-86cb917d18e6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.042417] env[61243]: DEBUG nova.network.neutron [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 676.213943] env[61243]: INFO nova.compute.manager [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 8efd391b-7b76-4167-bd18-48566271ccc8] Took 1.04 seconds to deallocate network for instance. [ 676.226718] env[61243]: DEBUG nova.compute.manager [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 676.600989] env[61243]: DEBUG nova.network.neutron [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.723795] env[61243]: DEBUG nova.network.neutron [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.757818] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.955874] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad017e10-11d1-412e-9f29-f6c6936c99f9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.964553] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-432359e1-333b-45c7-915e-9ac85492d60b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.996667] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22b8f3b3-6643-4ee2-be3d-0c71f5e872e4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.005222] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eeeb41f-609a-4237-ac8f-ad935730b606 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.019707] env[61243]: DEBUG nova.compute.provider_tree [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.226558] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Releasing lock "refresh_cache-84529438-8560-4323-a2f8-86cb917d18e6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.226558] env[61243]: DEBUG nova.compute.manager [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 677.226558] env[61243]: DEBUG nova.compute.manager [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 677.227225] env[61243]: DEBUG nova.network.neutron [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 677.256412] env[61243]: DEBUG nova.network.neutron [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.272097] env[61243]: INFO nova.scheduler.client.report [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Deleted allocations for instance 8efd391b-7b76-4167-bd18-48566271ccc8 [ 677.522987] env[61243]: DEBUG nova.scheduler.client.report [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.766035] env[61243]: DEBUG nova.network.neutron [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.781968] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e866e2ff-ec00-4b34-bb34-fa29480d3019 tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Lock "8efd391b-7b76-4167-bd18-48566271ccc8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.938s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.886946] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "f42f07cd-3d94-41a3-b202-ca1fa6b75da5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.886946] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "f42f07cd-3d94-41a3-b202-ca1fa6b75da5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.033951] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.997s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.034608] env[61243]: ERROR nova.compute.manager [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c1b660a5-8cfb-4650-9a8b-fa7bf0c491ab, please check neutron logs for more information. [ 678.034608] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Traceback (most recent call last): [ 678.034608] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 678.034608] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] self.driver.spawn(context, instance, image_meta, [ 678.034608] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 678.034608] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] self._vmops.spawn(context, instance, image_meta, injected_files, [ 678.034608] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 678.034608] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] vm_ref = self.build_virtual_machine(instance, [ 678.034608] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 678.034608] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] vif_infos = vmwarevif.get_vif_info(self._session, [ 678.034608] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 678.035015] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] for vif in network_info: [ 678.035015] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 678.035015] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] return self._sync_wrapper(fn, *args, **kwargs) [ 678.035015] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 678.035015] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] self.wait() [ 678.035015] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 678.035015] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] self[:] = self._gt.wait() [ 678.035015] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 678.035015] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] return self._exit_event.wait() [ 678.035015] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 678.035015] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] current.throw(*self._exc) [ 678.035015] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 678.035015] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] result = function(*args, **kwargs) [ 678.035400] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 678.035400] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] return func(*args, **kwargs) [ 678.035400] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 678.035400] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] raise e [ 678.035400] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.035400] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] nwinfo = self.network_api.allocate_for_instance( [ 678.035400] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 678.035400] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] created_port_ids = self._update_ports_for_instance( [ 678.035400] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 678.035400] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] with excutils.save_and_reraise_exception(): [ 678.035400] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.035400] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] self.force_reraise() [ 678.035400] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.036128] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] raise self.value [ 678.036128] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 678.036128] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] updated_port = self._update_port( [ 678.036128] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.036128] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] _ensure_no_port_binding_failure(port) [ 678.036128] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.036128] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] raise exception.PortBindingFailed(port_id=port['id']) [ 678.036128] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] nova.exception.PortBindingFailed: Binding failed for port c1b660a5-8cfb-4650-9a8b-fa7bf0c491ab, please check neutron logs for more information. [ 678.036128] env[61243]: ERROR nova.compute.manager [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] [ 678.036128] env[61243]: DEBUG nova.compute.utils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Binding failed for port c1b660a5-8cfb-4650-9a8b-fa7bf0c491ab, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 678.036644] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.670s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.041871] env[61243]: DEBUG nova.compute.manager [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Build of instance 4072d5a6-f69d-4fb9-837b-7855054e5206 was re-scheduled: Binding failed for port c1b660a5-8cfb-4650-9a8b-fa7bf0c491ab, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 678.042232] env[61243]: DEBUG nova.compute.manager [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 678.042298] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Acquiring lock "refresh_cache-4072d5a6-f69d-4fb9-837b-7855054e5206" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.043638] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Acquired lock "refresh_cache-4072d5a6-f69d-4fb9-837b-7855054e5206" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.043638] env[61243]: DEBUG nova.network.neutron [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 678.273848] env[61243]: INFO nova.compute.manager [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 84529438-8560-4323-a2f8-86cb917d18e6] Took 1.05 seconds to deallocate network for instance. [ 678.285688] env[61243]: DEBUG nova.compute.manager [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 678.566204] env[61243]: DEBUG nova.network.neutron [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.625866] env[61243]: DEBUG nova.network.neutron [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.816972] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.990611] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d20822-042d-4729-8699-ee52a7db4030 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.999439] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7d1f56-60f6-4b67-bfde-815238f817d7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.034298] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a097bdda-6de3-4ba1-8e9d-f9d8a2829e5e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.042044] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbdd01fa-8f0a-4938-ba98-30437c5d0d5e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.055686] env[61243]: DEBUG nova.compute.provider_tree [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.128763] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Releasing lock "refresh_cache-4072d5a6-f69d-4fb9-837b-7855054e5206" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.129096] env[61243]: DEBUG nova.compute.manager [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 679.129263] env[61243]: DEBUG nova.compute.manager [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 679.129457] env[61243]: DEBUG nova.network.neutron [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 679.146537] env[61243]: DEBUG nova.network.neutron [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.326384] env[61243]: INFO nova.scheduler.client.report [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Deleted allocations for instance 84529438-8560-4323-a2f8-86cb917d18e6 [ 679.480071] env[61243]: DEBUG oslo_concurrency.lockutils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Acquiring lock "6bd8ea50-e23f-48e7-9979-019e7a694c66" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.480071] env[61243]: DEBUG oslo_concurrency.lockutils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Lock "6bd8ea50-e23f-48e7-9979-019e7a694c66" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.559172] env[61243]: DEBUG nova.scheduler.client.report [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.651282] env[61243]: DEBUG nova.network.neutron [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.837870] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c3aae76e-1c5b-43be-805f-d065057d96b3 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Lock "84529438-8560-4323-a2f8-86cb917d18e6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.636s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.064972] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.028s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.065923] env[61243]: ERROR nova.compute.manager [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0d346e84-b8e6-4678-99ad-bf03d7a94af8, please check neutron logs for more information. [ 680.065923] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Traceback (most recent call last): [ 680.065923] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.065923] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] self.driver.spawn(context, instance, image_meta, [ 680.065923] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 680.065923] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.065923] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.065923] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] vm_ref = self.build_virtual_machine(instance, [ 680.065923] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.065923] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.065923] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.066232] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] for vif in network_info: [ 680.066232] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.066232] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] return self._sync_wrapper(fn, *args, **kwargs) [ 680.066232] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.066232] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] self.wait() [ 680.066232] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.066232] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] self[:] = self._gt.wait() [ 680.066232] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.066232] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] return self._exit_event.wait() [ 680.066232] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 680.066232] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] current.throw(*self._exc) [ 680.066232] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.066232] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] result = function(*args, **kwargs) [ 680.066557] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 680.066557] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] return func(*args, **kwargs) [ 680.066557] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.066557] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] raise e [ 680.066557] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.066557] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] nwinfo = self.network_api.allocate_for_instance( [ 680.066557] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.066557] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] created_port_ids = self._update_ports_for_instance( [ 680.066557] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.066557] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] with excutils.save_and_reraise_exception(): [ 680.066557] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.066557] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] self.force_reraise() [ 680.066557] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.066860] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] raise self.value [ 680.066860] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.066860] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] updated_port = self._update_port( [ 680.066860] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.066860] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] _ensure_no_port_binding_failure(port) [ 680.066860] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.066860] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] raise exception.PortBindingFailed(port_id=port['id']) [ 680.066860] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] nova.exception.PortBindingFailed: Binding failed for port 0d346e84-b8e6-4678-99ad-bf03d7a94af8, please check neutron logs for more information. [ 680.066860] env[61243]: ERROR nova.compute.manager [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] [ 680.066860] env[61243]: DEBUG nova.compute.utils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Binding failed for port 0d346e84-b8e6-4678-99ad-bf03d7a94af8, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 680.067583] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.560s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.069654] env[61243]: DEBUG nova.compute.manager [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Build of instance d113d9ef-5698-4851-a51a-7c7c807e9135 was re-scheduled: Binding failed for port 0d346e84-b8e6-4678-99ad-bf03d7a94af8, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 680.069874] env[61243]: DEBUG nova.compute.manager [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 680.070114] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Acquiring lock "refresh_cache-d113d9ef-5698-4851-a51a-7c7c807e9135" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.070260] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Acquired lock "refresh_cache-d113d9ef-5698-4851-a51a-7c7c807e9135" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.070428] env[61243]: DEBUG nova.network.neutron [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 680.154854] env[61243]: INFO nova.compute.manager [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] [instance: 4072d5a6-f69d-4fb9-837b-7855054e5206] Took 1.03 seconds to deallocate network for instance. [ 680.340631] env[61243]: DEBUG nova.compute.manager [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 680.604771] env[61243]: DEBUG nova.network.neutron [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.750082] env[61243]: DEBUG nova.network.neutron [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.877673] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.148636] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Acquiring lock "e9b2e5d1-e75e-4346-a0af-c0dafac0692e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.148819] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Lock "e9b2e5d1-e75e-4346-a0af-c0dafac0692e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.194106] env[61243]: INFO nova.scheduler.client.report [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Deleted allocations for instance 4072d5a6-f69d-4fb9-837b-7855054e5206 [ 681.251570] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Releasing lock "refresh_cache-d113d9ef-5698-4851-a51a-7c7c807e9135" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.251987] env[61243]: DEBUG nova.compute.manager [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 681.253014] env[61243]: DEBUG nova.compute.manager [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.253579] env[61243]: DEBUG nova.network.neutron [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 681.272559] env[61243]: DEBUG nova.network.neutron [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.614991] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 4072d5a6-f69d-4fb9-837b-7855054e5206 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 681.706201] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c12cc3b-db48-45ce-8d59-d6c5eb027787 tempest-ListImageFiltersTestJSON-1622113442 tempest-ListImageFiltersTestJSON-1622113442-project-member] Lock "4072d5a6-f69d-4fb9-837b-7855054e5206" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.990s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.779134] env[61243]: DEBUG nova.network.neutron [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.119330] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance d113d9ef-5698-4851-a51a-7c7c807e9135 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 682.119330] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 06264429-11eb-4d64-9f6b-6587cf048c41 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 682.119498] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance d5131a9f-f4bd-40c8-98d1-043651537a16 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 682.211571] env[61243]: DEBUG nova.compute.manager [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 682.283408] env[61243]: INFO nova.compute.manager [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: d113d9ef-5698-4851-a51a-7c7c807e9135] Took 1.03 seconds to deallocate network for instance. [ 682.361284] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "86179fb0-99df-4b10-a815-c19168e9521e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.361731] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "86179fb0-99df-4b10-a815-c19168e9521e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.623376] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 72aa2816-5e02-4dc2-9d6a-c3e58c651224 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 682.735804] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.130160] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance c887c01b-a67e-4e3e-9e17-1deb26e9b508 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 683.328693] env[61243]: INFO nova.scheduler.client.report [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Deleted allocations for instance d113d9ef-5698-4851-a51a-7c7c807e9135 [ 683.634231] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance cceaed50-e20f-45e4-9959-29dac2ecad30 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 683.842868] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8a077ca2-8b02-41a9-9759-45308c20d498 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Lock "d113d9ef-5698-4851-a51a-7c7c807e9135" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.844s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.138877] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 755cdc16-0eae-4978-917f-634a8ac7af8f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 684.345995] env[61243]: DEBUG nova.compute.manager [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 684.642051] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 37b43ee9-f991-4d56-8fcc-192f6534f803 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 684.878356] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.144900] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 8e85f285-1190-4c15-bc1a-8364434dc4ba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 685.648812] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance f6566092-9921-4a94-a0b5-a2d3803a9d18 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 686.152303] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance fd2d724a-76b6-405f-b003-54de2ad84f5a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 686.655359] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 3b852a19-9f86-486b-b98d-e87ffc54d943 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 687.159317] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance be8ac265-b214-4db5-9781-d171cee46538 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 687.353941] env[61243]: DEBUG oslo_concurrency.lockutils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "07ee984d-476e-484d-ba80-0ec2e411faa9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.354291] env[61243]: DEBUG oslo_concurrency.lockutils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.661405] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance e02f41be-125a-4228-8136-9e6639a268a9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.166025] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance b1bad82d-d7bf-4e58-bbf2-f3d68f214115 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.669299] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 689.173724] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 454d97a2-40d0-4c7a-9df7-04c391715e0d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 689.677056] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 2e25cb9a-9b21-4b60-b741-944e53bd50e9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 690.179144] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 4a0a08ca-bbf5-4124-853e-0be45f531d91 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 690.683018] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 691.185246] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 691.688900] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 1f4985a7-f6cc-4e79-aea3-ce0623aed515 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.192333] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 5314f922-6603-40dd-8760-9c1b77d6bd0a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.696074] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance f42f07cd-3d94-41a3-b202-ca1fa6b75da5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.199632] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 6bd8ea50-e23f-48e7-9979-019e7a694c66 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.702457] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance e9b2e5d1-e75e-4346-a0af-c0dafac0692e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.702792] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 693.702948] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 694.005798] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827d7211-f687-41cf-bcf7-7d8181fc220d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.013381] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b530c0b3-ca1f-4812-8d63-f2b7fcd1ac4f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.044332] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04c82d05-541f-4eff-a294-4e727a39737a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.051371] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-651c3928-1f9b-4b1a-9f77-c680d9f17b31 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.064358] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.568113] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 695.075880] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61243) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 695.076203] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 15.009s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.076473] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.312s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.078119] env[61243]: INFO nova.compute.claims [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 696.381556] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d67381-c962-456a-87f8-1566d71fd669 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.390015] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94248bfe-02de-4a61-93a1-a96119cf1259 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.419152] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58c03ba4-e25d-4132-a67a-69fc0c588f2f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.426132] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b27d5c-01ef-46cb-bfa1-91aa712fda4b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.438797] env[61243]: DEBUG nova.compute.provider_tree [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.942156] env[61243]: DEBUG nova.scheduler.client.report [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 697.448146] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.448710] env[61243]: DEBUG nova.compute.manager [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 697.451873] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.051s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.956445] env[61243]: DEBUG nova.compute.utils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 697.958192] env[61243]: DEBUG nova.compute.manager [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 697.958377] env[61243]: DEBUG nova.network.neutron [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 698.027211] env[61243]: DEBUG nova.policy [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c391b37eb72c4d25b47737372a7fa78d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '464cc6cb502047efbf4ec6bb80a0d310', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 698.288010] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69ea4f5d-13cc-4996-8f09-0367c02f32e8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.294660] env[61243]: DEBUG nova.network.neutron [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Successfully created port: 9d5a7a9b-a4c9-4217-b740-0c33d86109a2 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 698.298294] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1b227d1-f6af-4991-899c-389617230474 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.328905] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a8bdae-484d-4c38-bb3d-22c890801f97 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.335912] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f956f6a2-bbcc-471a-a13b-24d332a86134 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.348726] env[61243]: DEBUG nova.compute.provider_tree [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.464144] env[61243]: DEBUG nova.compute.manager [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 698.851856] env[61243]: DEBUG nova.scheduler.client.report [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 698.992227] env[61243]: DEBUG nova.compute.manager [req-1d97a1bd-b093-446e-ba84-98726c82c5c1 req-84ad98b5-fff0-4658-87ca-41999d0a6783 service nova] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Received event network-changed-9d5a7a9b-a4c9-4217-b740-0c33d86109a2 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 698.993029] env[61243]: DEBUG nova.compute.manager [req-1d97a1bd-b093-446e-ba84-98726c82c5c1 req-84ad98b5-fff0-4658-87ca-41999d0a6783 service nova] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Refreshing instance network info cache due to event network-changed-9d5a7a9b-a4c9-4217-b740-0c33d86109a2. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 698.993029] env[61243]: DEBUG oslo_concurrency.lockutils [req-1d97a1bd-b093-446e-ba84-98726c82c5c1 req-84ad98b5-fff0-4658-87ca-41999d0a6783 service nova] Acquiring lock "refresh_cache-72aa2816-5e02-4dc2-9d6a-c3e58c651224" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.993029] env[61243]: DEBUG oslo_concurrency.lockutils [req-1d97a1bd-b093-446e-ba84-98726c82c5c1 req-84ad98b5-fff0-4658-87ca-41999d0a6783 service nova] Acquired lock "refresh_cache-72aa2816-5e02-4dc2-9d6a-c3e58c651224" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.993029] env[61243]: DEBUG nova.network.neutron [req-1d97a1bd-b093-446e-ba84-98726c82c5c1 req-84ad98b5-fff0-4658-87ca-41999d0a6783 service nova] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Refreshing network info cache for port 9d5a7a9b-a4c9-4217-b740-0c33d86109a2 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 699.141418] env[61243]: ERROR nova.compute.manager [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9d5a7a9b-a4c9-4217-b740-0c33d86109a2, please check neutron logs for more information. [ 699.141418] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 699.141418] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.141418] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 699.141418] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.141418] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 699.141418] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.141418] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 699.141418] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.141418] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 699.141418] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.141418] env[61243]: ERROR nova.compute.manager raise self.value [ 699.141418] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.141418] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 699.141418] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.141418] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 699.142200] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.142200] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 699.142200] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9d5a7a9b-a4c9-4217-b740-0c33d86109a2, please check neutron logs for more information. [ 699.142200] env[61243]: ERROR nova.compute.manager [ 699.142200] env[61243]: Traceback (most recent call last): [ 699.142200] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 699.142200] env[61243]: listener.cb(fileno) [ 699.142200] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.142200] env[61243]: result = function(*args, **kwargs) [ 699.142200] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.142200] env[61243]: return func(*args, **kwargs) [ 699.142200] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.142200] env[61243]: raise e [ 699.142200] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.142200] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 699.142200] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.142200] env[61243]: created_port_ids = self._update_ports_for_instance( [ 699.142200] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.142200] env[61243]: with excutils.save_and_reraise_exception(): [ 699.142200] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.142200] env[61243]: self.force_reraise() [ 699.142200] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.142200] env[61243]: raise self.value [ 699.142200] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.142200] env[61243]: updated_port = self._update_port( [ 699.142200] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.142200] env[61243]: _ensure_no_port_binding_failure(port) [ 699.142200] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.142200] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 699.143374] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 9d5a7a9b-a4c9-4217-b740-0c33d86109a2, please check neutron logs for more information. [ 699.143374] env[61243]: Removing descriptor: 16 [ 699.357119] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.905s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.357790] env[61243]: ERROR nova.compute.manager [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 449658ee-44ec-497f-b8a3-54799b7ea061, please check neutron logs for more information. [ 699.357790] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Traceback (most recent call last): [ 699.357790] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.357790] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] self.driver.spawn(context, instance, image_meta, [ 699.357790] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 699.357790] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.357790] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.357790] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] vm_ref = self.build_virtual_machine(instance, [ 699.357790] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.357790] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.357790] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.358152] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] for vif in network_info: [ 699.358152] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.358152] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] return self._sync_wrapper(fn, *args, **kwargs) [ 699.358152] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.358152] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] self.wait() [ 699.358152] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.358152] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] self[:] = self._gt.wait() [ 699.358152] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.358152] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] return self._exit_event.wait() [ 699.358152] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 699.358152] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] current.throw(*self._exc) [ 699.358152] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.358152] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] result = function(*args, **kwargs) [ 699.358534] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.358534] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] return func(*args, **kwargs) [ 699.358534] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.358534] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] raise e [ 699.358534] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.358534] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] nwinfo = self.network_api.allocate_for_instance( [ 699.358534] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.358534] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] created_port_ids = self._update_ports_for_instance( [ 699.358534] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.358534] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] with excutils.save_and_reraise_exception(): [ 699.358534] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.358534] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] self.force_reraise() [ 699.358534] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.358842] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] raise self.value [ 699.358842] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.358842] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] updated_port = self._update_port( [ 699.358842] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.358842] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] _ensure_no_port_binding_failure(port) [ 699.358842] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.358842] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] raise exception.PortBindingFailed(port_id=port['id']) [ 699.358842] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] nova.exception.PortBindingFailed: Binding failed for port 449658ee-44ec-497f-b8a3-54799b7ea061, please check neutron logs for more information. [ 699.358842] env[61243]: ERROR nova.compute.manager [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] [ 699.358842] env[61243]: DEBUG nova.compute.utils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Binding failed for port 449658ee-44ec-497f-b8a3-54799b7ea061, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 699.359863] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.126s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.361381] env[61243]: INFO nova.compute.claims [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 699.364125] env[61243]: DEBUG nova.compute.manager [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Build of instance 06264429-11eb-4d64-9f6b-6587cf048c41 was re-scheduled: Binding failed for port 449658ee-44ec-497f-b8a3-54799b7ea061, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 699.364514] env[61243]: DEBUG nova.compute.manager [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 699.364742] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Acquiring lock "refresh_cache-06264429-11eb-4d64-9f6b-6587cf048c41" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.364892] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Acquired lock "refresh_cache-06264429-11eb-4d64-9f6b-6587cf048c41" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.365059] env[61243]: DEBUG nova.network.neutron [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 699.473212] env[61243]: DEBUG nova.compute.manager [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 699.500022] env[61243]: DEBUG nova.virt.hardware [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 699.500316] env[61243]: DEBUG nova.virt.hardware [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 699.500444] env[61243]: DEBUG nova.virt.hardware [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 699.500619] env[61243]: DEBUG nova.virt.hardware [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 699.500791] env[61243]: DEBUG nova.virt.hardware [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 699.500900] env[61243]: DEBUG nova.virt.hardware [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 699.501125] env[61243]: DEBUG nova.virt.hardware [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 699.501327] env[61243]: DEBUG nova.virt.hardware [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 699.501505] env[61243]: DEBUG nova.virt.hardware [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 699.501669] env[61243]: DEBUG nova.virt.hardware [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 699.501839] env[61243]: DEBUG nova.virt.hardware [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.502751] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff0444f-1e6f-48ed-b497-7bef46c26308 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.511037] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76cf4be3-d16b-4209-a5e5-21ac77699f70 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.525911] env[61243]: ERROR nova.compute.manager [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9d5a7a9b-a4c9-4217-b740-0c33d86109a2, please check neutron logs for more information. [ 699.525911] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Traceback (most recent call last): [ 699.525911] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 699.525911] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] yield resources [ 699.525911] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.525911] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] self.driver.spawn(context, instance, image_meta, [ 699.525911] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 699.525911] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.525911] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.525911] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] vm_ref = self.build_virtual_machine(instance, [ 699.525911] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.526430] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.526430] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.526430] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] for vif in network_info: [ 699.526430] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.526430] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] return self._sync_wrapper(fn, *args, **kwargs) [ 699.526430] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.526430] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] self.wait() [ 699.526430] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.526430] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] self[:] = self._gt.wait() [ 699.526430] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.526430] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] return self._exit_event.wait() [ 699.526430] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 699.526430] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] current.throw(*self._exc) [ 699.526833] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.526833] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] result = function(*args, **kwargs) [ 699.526833] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.526833] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] return func(*args, **kwargs) [ 699.526833] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.526833] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] raise e [ 699.526833] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.526833] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] nwinfo = self.network_api.allocate_for_instance( [ 699.526833] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.526833] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] created_port_ids = self._update_ports_for_instance( [ 699.526833] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.526833] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] with excutils.save_and_reraise_exception(): [ 699.526833] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.527191] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] self.force_reraise() [ 699.527191] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.527191] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] raise self.value [ 699.527191] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.527191] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] updated_port = self._update_port( [ 699.527191] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.527191] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] _ensure_no_port_binding_failure(port) [ 699.527191] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.527191] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] raise exception.PortBindingFailed(port_id=port['id']) [ 699.527191] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] nova.exception.PortBindingFailed: Binding failed for port 9d5a7a9b-a4c9-4217-b740-0c33d86109a2, please check neutron logs for more information. [ 699.527191] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] [ 699.527191] env[61243]: INFO nova.compute.manager [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Terminating instance [ 699.528248] env[61243]: DEBUG nova.network.neutron [req-1d97a1bd-b093-446e-ba84-98726c82c5c1 req-84ad98b5-fff0-4658-87ca-41999d0a6783 service nova] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.530280] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Acquiring lock "refresh_cache-72aa2816-5e02-4dc2-9d6a-c3e58c651224" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.603009] env[61243]: DEBUG nova.network.neutron [req-1d97a1bd-b093-446e-ba84-98726c82c5c1 req-84ad98b5-fff0-4658-87ca-41999d0a6783 service nova] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.885236] env[61243]: DEBUG nova.network.neutron [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.961646] env[61243]: DEBUG nova.network.neutron [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.106090] env[61243]: DEBUG oslo_concurrency.lockutils [req-1d97a1bd-b093-446e-ba84-98726c82c5c1 req-84ad98b5-fff0-4658-87ca-41999d0a6783 service nova] Releasing lock "refresh_cache-72aa2816-5e02-4dc2-9d6a-c3e58c651224" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.106948] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Acquired lock "refresh_cache-72aa2816-5e02-4dc2-9d6a-c3e58c651224" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.106948] env[61243]: DEBUG nova.network.neutron [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 700.465387] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Releasing lock "refresh_cache-06264429-11eb-4d64-9f6b-6587cf048c41" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.465387] env[61243]: DEBUG nova.compute.manager [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 700.465387] env[61243]: DEBUG nova.compute.manager [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 700.465387] env[61243]: DEBUG nova.network.neutron [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 700.481668] env[61243]: DEBUG nova.network.neutron [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.629025] env[61243]: DEBUG nova.network.neutron [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.666052] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2062034a-37f6-4d40-bf4c-93d9cdd0b637 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.674912] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1b7f46-fa5f-49df-868f-5fc921fdd92f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.705156] env[61243]: DEBUG nova.network.neutron [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.706746] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e18296-0efe-4496-85bb-57ea4dc3af57 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.714102] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c75615a-7417-499c-8256-40bb90832cd2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.727466] env[61243]: DEBUG nova.compute.provider_tree [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.984729] env[61243]: DEBUG nova.network.neutron [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.027776] env[61243]: DEBUG nova.compute.manager [req-b5cd10fc-0a88-4b3a-b279-d0266cf9e61e req-7f137be0-5983-430c-99b0-72780140fccd service nova] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Received event network-vif-deleted-9d5a7a9b-a4c9-4217-b740-0c33d86109a2 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 701.210298] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Releasing lock "refresh_cache-72aa2816-5e02-4dc2-9d6a-c3e58c651224" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.210688] env[61243]: DEBUG nova.compute.manager [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 701.210888] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 701.211260] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c298ae14-b1ca-4fb7-85a8-fe50ccabd549 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.220172] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8ddf54-c4eb-4dd1-a307-90c33527e1f4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.230839] env[61243]: DEBUG nova.scheduler.client.report [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.244986] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 72aa2816-5e02-4dc2-9d6a-c3e58c651224 could not be found. [ 701.246067] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 701.246067] env[61243]: INFO nova.compute.manager [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Took 0.03 seconds to destroy the instance on the hypervisor. [ 701.246067] env[61243]: DEBUG oslo.service.loopingcall [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 701.246426] env[61243]: DEBUG nova.compute.manager [-] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 701.246531] env[61243]: DEBUG nova.network.neutron [-] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 701.261855] env[61243]: DEBUG nova.network.neutron [-] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.487499] env[61243]: INFO nova.compute.manager [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] [instance: 06264429-11eb-4d64-9f6b-6587cf048c41] Took 1.02 seconds to deallocate network for instance. [ 701.736265] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.376s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.736790] env[61243]: DEBUG nova.compute.manager [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 701.739305] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.790s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.763827] env[61243]: DEBUG nova.network.neutron [-] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.244483] env[61243]: DEBUG nova.compute.utils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 702.248835] env[61243]: DEBUG nova.compute.manager [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 702.249018] env[61243]: DEBUG nova.network.neutron [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 702.265532] env[61243]: INFO nova.compute.manager [-] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Took 1.02 seconds to deallocate network for instance. [ 702.267766] env[61243]: DEBUG nova.compute.claims [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 702.267945] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.301381] env[61243]: DEBUG nova.policy [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5cdcbbd22199460da051f9604c31e9d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97fe5b9c92904b1c9992326adbbe8684', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 702.515030] env[61243]: INFO nova.scheduler.client.report [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Deleted allocations for instance 06264429-11eb-4d64-9f6b-6587cf048c41 [ 702.560346] env[61243]: DEBUG nova.network.neutron [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Successfully created port: bfc0ef48-c4b5-4571-8059-4253feb36455 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 702.574950] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fbb0ea5-66c1-45a6-80fd-a1542b597bf0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.582798] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0feccba3-4a3b-40f4-a172-0221ac743536 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.612781] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e301a668-4354-47c3-b3bd-3e9718f09bdd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.619928] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf28c1c-08f9-4e7b-bf73-8fa3ba02cc9b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.634086] env[61243]: DEBUG nova.compute.provider_tree [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.753849] env[61243]: DEBUG nova.compute.manager [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 703.026433] env[61243]: DEBUG oslo_concurrency.lockutils [None req-50ade4d0-a22a-4073-be96-25edbb7c0c78 tempest-ServersAdminTestJSON-560305871 tempest-ServersAdminTestJSON-560305871-project-member] Lock "06264429-11eb-4d64-9f6b-6587cf048c41" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.972s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.137648] env[61243]: DEBUG nova.scheduler.client.report [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.360441] env[61243]: DEBUG nova.compute.manager [req-530e90e8-847d-4d62-a8a2-23b4afccf142 req-c8d23c29-e470-49cd-8233-c8939d8bcbeb service nova] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Received event network-changed-bfc0ef48-c4b5-4571-8059-4253feb36455 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 703.360787] env[61243]: DEBUG nova.compute.manager [req-530e90e8-847d-4d62-a8a2-23b4afccf142 req-c8d23c29-e470-49cd-8233-c8939d8bcbeb service nova] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Refreshing instance network info cache due to event network-changed-bfc0ef48-c4b5-4571-8059-4253feb36455. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 703.360856] env[61243]: DEBUG oslo_concurrency.lockutils [req-530e90e8-847d-4d62-a8a2-23b4afccf142 req-c8d23c29-e470-49cd-8233-c8939d8bcbeb service nova] Acquiring lock "refresh_cache-c887c01b-a67e-4e3e-9e17-1deb26e9b508" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.360973] env[61243]: DEBUG oslo_concurrency.lockutils [req-530e90e8-847d-4d62-a8a2-23b4afccf142 req-c8d23c29-e470-49cd-8233-c8939d8bcbeb service nova] Acquired lock "refresh_cache-c887c01b-a67e-4e3e-9e17-1deb26e9b508" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.361155] env[61243]: DEBUG nova.network.neutron [req-530e90e8-847d-4d62-a8a2-23b4afccf142 req-c8d23c29-e470-49cd-8233-c8939d8bcbeb service nova] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Refreshing network info cache for port bfc0ef48-c4b5-4571-8059-4253feb36455 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 703.414793] env[61243]: ERROR nova.compute.manager [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bfc0ef48-c4b5-4571-8059-4253feb36455, please check neutron logs for more information. [ 703.414793] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 703.414793] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.414793] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 703.414793] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.414793] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 703.414793] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.414793] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 703.414793] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.414793] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 703.414793] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.414793] env[61243]: ERROR nova.compute.manager raise self.value [ 703.414793] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.414793] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 703.414793] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.414793] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 703.415257] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.415257] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 703.415257] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bfc0ef48-c4b5-4571-8059-4253feb36455, please check neutron logs for more information. [ 703.415257] env[61243]: ERROR nova.compute.manager [ 703.415257] env[61243]: Traceback (most recent call last): [ 703.415257] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 703.415257] env[61243]: listener.cb(fileno) [ 703.415257] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.415257] env[61243]: result = function(*args, **kwargs) [ 703.415257] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 703.415257] env[61243]: return func(*args, **kwargs) [ 703.415257] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.415257] env[61243]: raise e [ 703.415257] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.415257] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 703.415257] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.415257] env[61243]: created_port_ids = self._update_ports_for_instance( [ 703.415257] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.415257] env[61243]: with excutils.save_and_reraise_exception(): [ 703.415257] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.415257] env[61243]: self.force_reraise() [ 703.415257] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.415257] env[61243]: raise self.value [ 703.415257] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.415257] env[61243]: updated_port = self._update_port( [ 703.415257] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.415257] env[61243]: _ensure_no_port_binding_failure(port) [ 703.415257] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.415257] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 703.416131] env[61243]: nova.exception.PortBindingFailed: Binding failed for port bfc0ef48-c4b5-4571-8059-4253feb36455, please check neutron logs for more information. [ 703.416131] env[61243]: Removing descriptor: 16 [ 703.530023] env[61243]: DEBUG nova.compute.manager [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 703.642369] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.903s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.642872] env[61243]: ERROR nova.compute.manager [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6fb89631-00db-453d-999a-a666fbd00063, please check neutron logs for more information. [ 703.642872] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Traceback (most recent call last): [ 703.642872] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 703.642872] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] self.driver.spawn(context, instance, image_meta, [ 703.642872] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 703.642872] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.642872] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.642872] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] vm_ref = self.build_virtual_machine(instance, [ 703.642872] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.642872] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.642872] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.643175] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] for vif in network_info: [ 703.643175] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 703.643175] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] return self._sync_wrapper(fn, *args, **kwargs) [ 703.643175] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 703.643175] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] self.wait() [ 703.643175] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 703.643175] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] self[:] = self._gt.wait() [ 703.643175] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.643175] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] return self._exit_event.wait() [ 703.643175] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 703.643175] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] current.throw(*self._exc) [ 703.643175] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.643175] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] result = function(*args, **kwargs) [ 703.643490] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 703.643490] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] return func(*args, **kwargs) [ 703.643490] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.643490] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] raise e [ 703.643490] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.643490] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] nwinfo = self.network_api.allocate_for_instance( [ 703.643490] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.643490] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] created_port_ids = self._update_ports_for_instance( [ 703.643490] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.643490] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] with excutils.save_and_reraise_exception(): [ 703.643490] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.643490] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] self.force_reraise() [ 703.643490] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.643876] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] raise self.value [ 703.643876] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.643876] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] updated_port = self._update_port( [ 703.643876] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.643876] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] _ensure_no_port_binding_failure(port) [ 703.643876] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.643876] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] raise exception.PortBindingFailed(port_id=port['id']) [ 703.643876] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] nova.exception.PortBindingFailed: Binding failed for port 6fb89631-00db-453d-999a-a666fbd00063, please check neutron logs for more information. [ 703.643876] env[61243]: ERROR nova.compute.manager [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] [ 703.643876] env[61243]: DEBUG nova.compute.utils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Binding failed for port 6fb89631-00db-453d-999a-a666fbd00063, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 703.645346] env[61243]: DEBUG nova.compute.manager [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Build of instance d5131a9f-f4bd-40c8-98d1-043651537a16 was re-scheduled: Binding failed for port 6fb89631-00db-453d-999a-a666fbd00063, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 703.645779] env[61243]: DEBUG nova.compute.manager [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 703.646067] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "refresh_cache-d5131a9f-f4bd-40c8-98d1-043651537a16" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.646180] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired lock "refresh_cache-d5131a9f-f4bd-40c8-98d1-043651537a16" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.646345] env[61243]: DEBUG nova.network.neutron [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 703.647334] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.884s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.649282] env[61243]: INFO nova.compute.claims [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 703.764140] env[61243]: DEBUG nova.compute.manager [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 703.788149] env[61243]: DEBUG nova.virt.hardware [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 703.788397] env[61243]: DEBUG nova.virt.hardware [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 703.788555] env[61243]: DEBUG nova.virt.hardware [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 703.788780] env[61243]: DEBUG nova.virt.hardware [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 703.788878] env[61243]: DEBUG nova.virt.hardware [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 703.789040] env[61243]: DEBUG nova.virt.hardware [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 703.789256] env[61243]: DEBUG nova.virt.hardware [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 703.789439] env[61243]: DEBUG nova.virt.hardware [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 703.789609] env[61243]: DEBUG nova.virt.hardware [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 703.789779] env[61243]: DEBUG nova.virt.hardware [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 703.789957] env[61243]: DEBUG nova.virt.hardware [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 703.790805] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8158529-5da5-4563-a68c-ec425bce9d34 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.799165] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05613312-8605-47fd-bbd3-690fe28b5ef3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.814659] env[61243]: ERROR nova.compute.manager [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bfc0ef48-c4b5-4571-8059-4253feb36455, please check neutron logs for more information. [ 703.814659] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Traceback (most recent call last): [ 703.814659] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 703.814659] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] yield resources [ 703.814659] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 703.814659] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] self.driver.spawn(context, instance, image_meta, [ 703.814659] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 703.814659] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.814659] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.814659] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] vm_ref = self.build_virtual_machine(instance, [ 703.814659] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.815015] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.815015] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.815015] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] for vif in network_info: [ 703.815015] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 703.815015] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] return self._sync_wrapper(fn, *args, **kwargs) [ 703.815015] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 703.815015] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] self.wait() [ 703.815015] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 703.815015] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] self[:] = self._gt.wait() [ 703.815015] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.815015] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] return self._exit_event.wait() [ 703.815015] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 703.815015] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] current.throw(*self._exc) [ 703.815343] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.815343] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] result = function(*args, **kwargs) [ 703.815343] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 703.815343] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] return func(*args, **kwargs) [ 703.815343] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.815343] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] raise e [ 703.815343] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.815343] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] nwinfo = self.network_api.allocate_for_instance( [ 703.815343] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.815343] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] created_port_ids = self._update_ports_for_instance( [ 703.815343] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.815343] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] with excutils.save_and_reraise_exception(): [ 703.815343] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.815909] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] self.force_reraise() [ 703.815909] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.815909] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] raise self.value [ 703.815909] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.815909] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] updated_port = self._update_port( [ 703.815909] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.815909] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] _ensure_no_port_binding_failure(port) [ 703.815909] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.815909] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] raise exception.PortBindingFailed(port_id=port['id']) [ 703.815909] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] nova.exception.PortBindingFailed: Binding failed for port bfc0ef48-c4b5-4571-8059-4253feb36455, please check neutron logs for more information. [ 703.815909] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] [ 703.815909] env[61243]: INFO nova.compute.manager [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Terminating instance [ 703.816992] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Acquiring lock "refresh_cache-c887c01b-a67e-4e3e-9e17-1deb26e9b508" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.901230] env[61243]: DEBUG nova.network.neutron [req-530e90e8-847d-4d62-a8a2-23b4afccf142 req-c8d23c29-e470-49cd-8233-c8939d8bcbeb service nova] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.036457] env[61243]: DEBUG nova.network.neutron [req-530e90e8-847d-4d62-a8a2-23b4afccf142 req-c8d23c29-e470-49cd-8233-c8939d8bcbeb service nova] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.054792] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.172567] env[61243]: DEBUG nova.network.neutron [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.289897] env[61243]: DEBUG nova.network.neutron [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.539380] env[61243]: DEBUG oslo_concurrency.lockutils [req-530e90e8-847d-4d62-a8a2-23b4afccf142 req-c8d23c29-e470-49cd-8233-c8939d8bcbeb service nova] Releasing lock "refresh_cache-c887c01b-a67e-4e3e-9e17-1deb26e9b508" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.539846] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Acquired lock "refresh_cache-c887c01b-a67e-4e3e-9e17-1deb26e9b508" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.540051] env[61243]: DEBUG nova.network.neutron [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 704.793334] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Releasing lock "refresh_cache-d5131a9f-f4bd-40c8-98d1-043651537a16" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.793610] env[61243]: DEBUG nova.compute.manager [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 704.793801] env[61243]: DEBUG nova.compute.manager [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 704.793970] env[61243]: DEBUG nova.network.neutron [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 704.812725] env[61243]: DEBUG nova.network.neutron [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.970031] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ffc474d-fcd6-4ebf-9c58-92b0bfc653fa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.978197] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a53768b-5fa2-4f6f-bf6c-f9f3561bf134 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.007656] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a135dd1e-5a13-459a-b8c8-9b55c93a33f3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.014766] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6a83b2-6910-410e-9a94-b25ce7167d29 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.027839] env[61243]: DEBUG nova.compute.provider_tree [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.063745] env[61243]: DEBUG nova.network.neutron [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.189098] env[61243]: DEBUG nova.network.neutron [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.315271] env[61243]: DEBUG nova.network.neutron [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.391102] env[61243]: DEBUG nova.compute.manager [req-4b1baf52-735b-4af2-a303-767c06f16946 req-64ea8d90-1281-401f-a806-ae915fda9fb4 service nova] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Received event network-vif-deleted-bfc0ef48-c4b5-4571-8059-4253feb36455 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 705.531424] env[61243]: DEBUG nova.scheduler.client.report [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.693990] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Releasing lock "refresh_cache-c887c01b-a67e-4e3e-9e17-1deb26e9b508" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.694392] env[61243]: DEBUG nova.compute.manager [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 705.694586] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 705.694893] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a83c61b7-2059-4f92-96e8-ed38595d1ff7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.704114] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb5da8e0-dee1-4695-91e6-bc343d46180f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.725166] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c887c01b-a67e-4e3e-9e17-1deb26e9b508 could not be found. [ 705.725400] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 705.725586] env[61243]: INFO nova.compute.manager [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Took 0.03 seconds to destroy the instance on the hypervisor. [ 705.725829] env[61243]: DEBUG oslo.service.loopingcall [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 705.726062] env[61243]: DEBUG nova.compute.manager [-] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 705.726162] env[61243]: DEBUG nova.network.neutron [-] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 705.745555] env[61243]: DEBUG nova.network.neutron [-] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.817818] env[61243]: INFO nova.compute.manager [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: d5131a9f-f4bd-40c8-98d1-043651537a16] Took 1.02 seconds to deallocate network for instance. [ 706.040215] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.390s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.040215] env[61243]: DEBUG nova.compute.manager [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 706.041809] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.284s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.043768] env[61243]: INFO nova.compute.claims [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 706.248027] env[61243]: DEBUG nova.network.neutron [-] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.553140] env[61243]: DEBUG nova.compute.utils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 706.557996] env[61243]: DEBUG nova.compute.manager [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 706.557996] env[61243]: DEBUG nova.network.neutron [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 706.604395] env[61243]: DEBUG nova.policy [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a40ade36ec3342cd8caec0de0028466b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a4e5047db6104fb59017304f96c467be', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 706.751616] env[61243]: INFO nova.compute.manager [-] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Took 1.03 seconds to deallocate network for instance. [ 706.756890] env[61243]: DEBUG nova.compute.claims [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 706.757125] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.849947] env[61243]: INFO nova.scheduler.client.report [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Deleted allocations for instance d5131a9f-f4bd-40c8-98d1-043651537a16 [ 706.894604] env[61243]: DEBUG nova.network.neutron [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Successfully created port: 88023220-3fa7-47f6-bb81-ff5e68503d8c {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 707.060986] env[61243]: DEBUG nova.compute.manager [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 707.358087] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6b6e40ed-fca4-4d46-be23-1523311bc32e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "d5131a9f-f4bd-40c8-98d1-043651537a16" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.290s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.410207] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e650846-bb3b-4f23-b3fa-c696e0a6562b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.417721] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2354d1-a96b-4ecb-9bee-66e132d5a9e0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.447744] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f931f524-12b4-499b-a549-2850b39af16e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.456029] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2536de53-32b1-4fc5-977d-a74516fdd4a3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.469191] env[61243]: DEBUG nova.compute.provider_tree [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.653402] env[61243]: DEBUG nova.compute.manager [req-574d7fcd-4086-40f4-b019-230612e57e5d req-1027d46d-8a79-4d05-bf33-216ee92a66e1 service nova] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Received event network-changed-88023220-3fa7-47f6-bb81-ff5e68503d8c {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 707.653587] env[61243]: DEBUG nova.compute.manager [req-574d7fcd-4086-40f4-b019-230612e57e5d req-1027d46d-8a79-4d05-bf33-216ee92a66e1 service nova] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Refreshing instance network info cache due to event network-changed-88023220-3fa7-47f6-bb81-ff5e68503d8c. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 707.653818] env[61243]: DEBUG oslo_concurrency.lockutils [req-574d7fcd-4086-40f4-b019-230612e57e5d req-1027d46d-8a79-4d05-bf33-216ee92a66e1 service nova] Acquiring lock "refresh_cache-cceaed50-e20f-45e4-9959-29dac2ecad30" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.653940] env[61243]: DEBUG oslo_concurrency.lockutils [req-574d7fcd-4086-40f4-b019-230612e57e5d req-1027d46d-8a79-4d05-bf33-216ee92a66e1 service nova] Acquired lock "refresh_cache-cceaed50-e20f-45e4-9959-29dac2ecad30" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.654677] env[61243]: DEBUG nova.network.neutron [req-574d7fcd-4086-40f4-b019-230612e57e5d req-1027d46d-8a79-4d05-bf33-216ee92a66e1 service nova] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Refreshing network info cache for port 88023220-3fa7-47f6-bb81-ff5e68503d8c {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 707.813432] env[61243]: ERROR nova.compute.manager [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 88023220-3fa7-47f6-bb81-ff5e68503d8c, please check neutron logs for more information. [ 707.813432] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 707.813432] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.813432] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 707.813432] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.813432] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 707.813432] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.813432] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 707.813432] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.813432] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 707.813432] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.813432] env[61243]: ERROR nova.compute.manager raise self.value [ 707.813432] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.813432] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 707.813432] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.813432] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 707.814134] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.814134] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 707.814134] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 88023220-3fa7-47f6-bb81-ff5e68503d8c, please check neutron logs for more information. [ 707.814134] env[61243]: ERROR nova.compute.manager [ 707.814134] env[61243]: Traceback (most recent call last): [ 707.814134] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 707.814134] env[61243]: listener.cb(fileno) [ 707.814134] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.814134] env[61243]: result = function(*args, **kwargs) [ 707.814134] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 707.814134] env[61243]: return func(*args, **kwargs) [ 707.814134] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.814134] env[61243]: raise e [ 707.814134] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.814134] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 707.814134] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.814134] env[61243]: created_port_ids = self._update_ports_for_instance( [ 707.814134] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.814134] env[61243]: with excutils.save_and_reraise_exception(): [ 707.814134] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.814134] env[61243]: self.force_reraise() [ 707.814134] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.814134] env[61243]: raise self.value [ 707.814134] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.814134] env[61243]: updated_port = self._update_port( [ 707.814134] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.814134] env[61243]: _ensure_no_port_binding_failure(port) [ 707.814134] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.814134] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 707.815168] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 88023220-3fa7-47f6-bb81-ff5e68503d8c, please check neutron logs for more information. [ 707.815168] env[61243]: Removing descriptor: 16 [ 707.860391] env[61243]: DEBUG nova.compute.manager [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 707.974030] env[61243]: DEBUG nova.scheduler.client.report [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.072065] env[61243]: DEBUG nova.compute.manager [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 708.099534] env[61243]: DEBUG nova.virt.hardware [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 708.099805] env[61243]: DEBUG nova.virt.hardware [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 708.099964] env[61243]: DEBUG nova.virt.hardware [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 708.100159] env[61243]: DEBUG nova.virt.hardware [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 708.100559] env[61243]: DEBUG nova.virt.hardware [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 708.100638] env[61243]: DEBUG nova.virt.hardware [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 708.101012] env[61243]: DEBUG nova.virt.hardware [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 708.101187] env[61243]: DEBUG nova.virt.hardware [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 708.101379] env[61243]: DEBUG nova.virt.hardware [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 708.101557] env[61243]: DEBUG nova.virt.hardware [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 708.101737] env[61243]: DEBUG nova.virt.hardware [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 708.102639] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a027f7-5816-44b3-82ef-1470e45a609d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.112541] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05123bb1-b107-4882-854d-59a19db38ffc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.129496] env[61243]: ERROR nova.compute.manager [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 88023220-3fa7-47f6-bb81-ff5e68503d8c, please check neutron logs for more information. [ 708.129496] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Traceback (most recent call last): [ 708.129496] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 708.129496] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] yield resources [ 708.129496] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 708.129496] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] self.driver.spawn(context, instance, image_meta, [ 708.129496] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 708.129496] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] self._vmops.spawn(context, instance, image_meta, injected_files, [ 708.129496] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 708.129496] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] vm_ref = self.build_virtual_machine(instance, [ 708.129496] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 708.130214] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] vif_infos = vmwarevif.get_vif_info(self._session, [ 708.130214] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 708.130214] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] for vif in network_info: [ 708.130214] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 708.130214] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] return self._sync_wrapper(fn, *args, **kwargs) [ 708.130214] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 708.130214] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] self.wait() [ 708.130214] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 708.130214] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] self[:] = self._gt.wait() [ 708.130214] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 708.130214] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] return self._exit_event.wait() [ 708.130214] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 708.130214] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] current.throw(*self._exc) [ 708.130878] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.130878] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] result = function(*args, **kwargs) [ 708.130878] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 708.130878] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] return func(*args, **kwargs) [ 708.130878] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.130878] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] raise e [ 708.130878] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.130878] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] nwinfo = self.network_api.allocate_for_instance( [ 708.130878] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 708.130878] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] created_port_ids = self._update_ports_for_instance( [ 708.130878] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 708.130878] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] with excutils.save_and_reraise_exception(): [ 708.130878] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.132316] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] self.force_reraise() [ 708.132316] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.132316] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] raise self.value [ 708.132316] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 708.132316] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] updated_port = self._update_port( [ 708.132316] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.132316] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] _ensure_no_port_binding_failure(port) [ 708.132316] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.132316] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] raise exception.PortBindingFailed(port_id=port['id']) [ 708.132316] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] nova.exception.PortBindingFailed: Binding failed for port 88023220-3fa7-47f6-bb81-ff5e68503d8c, please check neutron logs for more information. [ 708.132316] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] [ 708.132316] env[61243]: INFO nova.compute.manager [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Terminating instance [ 708.133265] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Acquiring lock "refresh_cache-cceaed50-e20f-45e4-9959-29dac2ecad30" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.175203] env[61243]: DEBUG nova.network.neutron [req-574d7fcd-4086-40f4-b019-230612e57e5d req-1027d46d-8a79-4d05-bf33-216ee92a66e1 service nova] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.254156] env[61243]: DEBUG nova.network.neutron [req-574d7fcd-4086-40f4-b019-230612e57e5d req-1027d46d-8a79-4d05-bf33-216ee92a66e1 service nova] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.384992] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.479149] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.481476] env[61243]: DEBUG nova.compute.manager [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 708.483419] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.667s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.485764] env[61243]: INFO nova.compute.claims [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 708.756773] env[61243]: DEBUG oslo_concurrency.lockutils [req-574d7fcd-4086-40f4-b019-230612e57e5d req-1027d46d-8a79-4d05-bf33-216ee92a66e1 service nova] Releasing lock "refresh_cache-cceaed50-e20f-45e4-9959-29dac2ecad30" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.757112] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Acquired lock "refresh_cache-cceaed50-e20f-45e4-9959-29dac2ecad30" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.757293] env[61243]: DEBUG nova.network.neutron [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 708.991214] env[61243]: DEBUG nova.compute.utils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 708.995150] env[61243]: DEBUG nova.compute.manager [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 708.995150] env[61243]: DEBUG nova.network.neutron [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 709.057237] env[61243]: DEBUG nova.policy [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c98c7ad8df4042a5aae96380c2dbb8ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d1226c068904a19a9a09c96e86c0567', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 709.084912] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "36e55334-8628-4dd7-a845-f4ae3d8e7ff9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.085167] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "36e55334-8628-4dd7-a845-f4ae3d8e7ff9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.335038] env[61243]: DEBUG nova.network.neutron [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.451243] env[61243]: DEBUG nova.network.neutron [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.495394] env[61243]: DEBUG nova.compute.manager [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 709.546906] env[61243]: DEBUG nova.network.neutron [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Successfully created port: 304f43ea-8b5b-46a6-9a7c-a355d8f4e546 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 709.717984] env[61243]: DEBUG nova.compute.manager [req-12b5fbba-fdbb-40a2-80fb-f728bb61c987 req-83d6500a-470f-4880-a300-1d3feb32e610 service nova] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Received event network-vif-deleted-88023220-3fa7-47f6-bb81-ff5e68503d8c {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 709.865875] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365694de-b3ba-4c3c-9a53-d369f5b10872 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.875448] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c3d2d5-cb9f-4ead-a6b8-8767d845ed7a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.917190] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f90533-cfc3-4c5c-808c-9e641707f637 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.925394] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-483d4e13-cb90-4167-85d1-9e1078e4f2a6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.940229] env[61243]: DEBUG nova.compute.provider_tree [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.953033] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Releasing lock "refresh_cache-cceaed50-e20f-45e4-9959-29dac2ecad30" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.953561] env[61243]: DEBUG nova.compute.manager [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 709.953767] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 709.954288] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-18a7b139-cbee-402e-9bc7-a48351b3f089 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.962383] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b8858d3-c624-4dd0-958b-44628eb943a8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.984555] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cceaed50-e20f-45e4-9959-29dac2ecad30 could not be found. [ 709.984898] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 709.985142] env[61243]: INFO nova.compute.manager [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Took 0.03 seconds to destroy the instance on the hypervisor. [ 709.985392] env[61243]: DEBUG oslo.service.loopingcall [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 709.985611] env[61243]: DEBUG nova.compute.manager [-] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 709.985708] env[61243]: DEBUG nova.network.neutron [-] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 710.002166] env[61243]: DEBUG nova.network.neutron [-] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.422680] env[61243]: ERROR nova.compute.manager [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 304f43ea-8b5b-46a6-9a7c-a355d8f4e546, please check neutron logs for more information. [ 710.422680] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 710.422680] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.422680] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 710.422680] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.422680] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 710.422680] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.422680] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 710.422680] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.422680] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 710.422680] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.422680] env[61243]: ERROR nova.compute.manager raise self.value [ 710.422680] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.422680] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 710.422680] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.422680] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 710.423130] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.423130] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 710.423130] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 304f43ea-8b5b-46a6-9a7c-a355d8f4e546, please check neutron logs for more information. [ 710.423130] env[61243]: ERROR nova.compute.manager [ 710.423130] env[61243]: Traceback (most recent call last): [ 710.423130] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 710.423130] env[61243]: listener.cb(fileno) [ 710.423130] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.423130] env[61243]: result = function(*args, **kwargs) [ 710.423130] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 710.423130] env[61243]: return func(*args, **kwargs) [ 710.423130] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.423130] env[61243]: raise e [ 710.423130] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.423130] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 710.423130] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.423130] env[61243]: created_port_ids = self._update_ports_for_instance( [ 710.423130] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.423130] env[61243]: with excutils.save_and_reraise_exception(): [ 710.423130] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.423130] env[61243]: self.force_reraise() [ 710.423130] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.423130] env[61243]: raise self.value [ 710.423130] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.423130] env[61243]: updated_port = self._update_port( [ 710.423130] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.423130] env[61243]: _ensure_no_port_binding_failure(port) [ 710.423130] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.423130] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 710.424031] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 304f43ea-8b5b-46a6-9a7c-a355d8f4e546, please check neutron logs for more information. [ 710.424031] env[61243]: Removing descriptor: 16 [ 710.444192] env[61243]: DEBUG nova.scheduler.client.report [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.507421] env[61243]: DEBUG nova.network.neutron [-] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.511209] env[61243]: DEBUG nova.compute.manager [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 710.536545] env[61243]: DEBUG nova.virt.hardware [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 710.536817] env[61243]: DEBUG nova.virt.hardware [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 710.537022] env[61243]: DEBUG nova.virt.hardware [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 710.537247] env[61243]: DEBUG nova.virt.hardware [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 710.537435] env[61243]: DEBUG nova.virt.hardware [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 710.537603] env[61243]: DEBUG nova.virt.hardware [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 710.537841] env[61243]: DEBUG nova.virt.hardware [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 710.538042] env[61243]: DEBUG nova.virt.hardware [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 710.538247] env[61243]: DEBUG nova.virt.hardware [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 710.538447] env[61243]: DEBUG nova.virt.hardware [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 710.538659] env[61243]: DEBUG nova.virt.hardware [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 710.539584] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86dcecec-127c-4250-8076-3b09db00a329 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.549724] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b90823-1bfe-427f-85a9-4b3ac8537ace {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.564050] env[61243]: ERROR nova.compute.manager [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 304f43ea-8b5b-46a6-9a7c-a355d8f4e546, please check neutron logs for more information. [ 710.564050] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Traceback (most recent call last): [ 710.564050] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 710.564050] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] yield resources [ 710.564050] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 710.564050] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] self.driver.spawn(context, instance, image_meta, [ 710.564050] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 710.564050] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.564050] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.564050] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] vm_ref = self.build_virtual_machine(instance, [ 710.564050] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.564462] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.564462] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.564462] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] for vif in network_info: [ 710.564462] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 710.564462] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] return self._sync_wrapper(fn, *args, **kwargs) [ 710.564462] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 710.564462] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] self.wait() [ 710.564462] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 710.564462] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] self[:] = self._gt.wait() [ 710.564462] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.564462] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] return self._exit_event.wait() [ 710.564462] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 710.564462] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] current.throw(*self._exc) [ 710.564863] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.564863] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] result = function(*args, **kwargs) [ 710.564863] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 710.564863] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] return func(*args, **kwargs) [ 710.564863] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.564863] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] raise e [ 710.564863] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.564863] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] nwinfo = self.network_api.allocate_for_instance( [ 710.564863] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.564863] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] created_port_ids = self._update_ports_for_instance( [ 710.564863] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.564863] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] with excutils.save_and_reraise_exception(): [ 710.564863] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.565287] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] self.force_reraise() [ 710.565287] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.565287] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] raise self.value [ 710.565287] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.565287] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] updated_port = self._update_port( [ 710.565287] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.565287] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] _ensure_no_port_binding_failure(port) [ 710.565287] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.565287] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] raise exception.PortBindingFailed(port_id=port['id']) [ 710.565287] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] nova.exception.PortBindingFailed: Binding failed for port 304f43ea-8b5b-46a6-9a7c-a355d8f4e546, please check neutron logs for more information. [ 710.565287] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] [ 710.565287] env[61243]: INFO nova.compute.manager [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Terminating instance [ 710.566575] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Acquiring lock "refresh_cache-755cdc16-0eae-4978-917f-634a8ac7af8f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.566769] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Acquired lock "refresh_cache-755cdc16-0eae-4978-917f-634a8ac7af8f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.566974] env[61243]: DEBUG nova.network.neutron [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 710.950154] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.465s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.950480] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.073s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.952048] env[61243]: INFO nova.compute.claims [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 711.009891] env[61243]: INFO nova.compute.manager [-] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Took 1.02 seconds to deallocate network for instance. [ 711.012661] env[61243]: DEBUG nova.compute.claims [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 711.012855] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.084022] env[61243]: DEBUG nova.network.neutron [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.135731] env[61243]: DEBUG nova.network.neutron [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.456737] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Acquiring lock "4edccecf-ef45-4c8b-8d24-19e596bd6482" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.456737] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Lock "4edccecf-ef45-4c8b-8d24-19e596bd6482" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.639066] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Releasing lock "refresh_cache-755cdc16-0eae-4978-917f-634a8ac7af8f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.639240] env[61243]: DEBUG nova.compute.manager [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 711.639423] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 711.639766] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ddeb7531-5032-40f3-abd3-378d0bf3cb5e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.648420] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2864f7b4-3f96-4a51-9bd9-2ef1811fd4a8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.669017] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 755cdc16-0eae-4978-917f-634a8ac7af8f could not be found. [ 711.669238] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 711.669458] env[61243]: INFO nova.compute.manager [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 711.669741] env[61243]: DEBUG oslo.service.loopingcall [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 711.669952] env[61243]: DEBUG nova.compute.manager [-] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 711.670062] env[61243]: DEBUG nova.network.neutron [-] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 711.685428] env[61243]: DEBUG nova.network.neutron [-] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.750509] env[61243]: DEBUG nova.compute.manager [req-6bad3402-a351-491b-92f8-0eeac98070d6 req-86b7b5ae-923a-44ec-b5e3-96f40a9ca3b1 service nova] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Received event network-changed-304f43ea-8b5b-46a6-9a7c-a355d8f4e546 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 711.750689] env[61243]: DEBUG nova.compute.manager [req-6bad3402-a351-491b-92f8-0eeac98070d6 req-86b7b5ae-923a-44ec-b5e3-96f40a9ca3b1 service nova] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Refreshing instance network info cache due to event network-changed-304f43ea-8b5b-46a6-9a7c-a355d8f4e546. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 711.750875] env[61243]: DEBUG oslo_concurrency.lockutils [req-6bad3402-a351-491b-92f8-0eeac98070d6 req-86b7b5ae-923a-44ec-b5e3-96f40a9ca3b1 service nova] Acquiring lock "refresh_cache-755cdc16-0eae-4978-917f-634a8ac7af8f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.750976] env[61243]: DEBUG oslo_concurrency.lockutils [req-6bad3402-a351-491b-92f8-0eeac98070d6 req-86b7b5ae-923a-44ec-b5e3-96f40a9ca3b1 service nova] Acquired lock "refresh_cache-755cdc16-0eae-4978-917f-634a8ac7af8f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.751146] env[61243]: DEBUG nova.network.neutron [req-6bad3402-a351-491b-92f8-0eeac98070d6 req-86b7b5ae-923a-44ec-b5e3-96f40a9ca3b1 service nova] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Refreshing network info cache for port 304f43ea-8b5b-46a6-9a7c-a355d8f4e546 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 711.960550] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Lock "4edccecf-ef45-4c8b-8d24-19e596bd6482" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.961191] env[61243]: DEBUG nova.compute.manager [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 712.188061] env[61243]: DEBUG nova.network.neutron [-] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.249311] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e4faf59-4c9e-4c5d-ad43-c237420f0a56 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.258145] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e7627f-7852-42f1-bef0-5c8546622b9b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.287400] env[61243]: DEBUG nova.network.neutron [req-6bad3402-a351-491b-92f8-0eeac98070d6 req-86b7b5ae-923a-44ec-b5e3-96f40a9ca3b1 service nova] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.290175] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4cd675-731c-40f1-9121-ebf15250fe61 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.296477] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3089382-10e5-4ed8-83b9-7ff61ccd145e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.309861] env[61243]: DEBUG nova.compute.provider_tree [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.342657] env[61243]: DEBUG nova.network.neutron [req-6bad3402-a351-491b-92f8-0eeac98070d6 req-86b7b5ae-923a-44ec-b5e3-96f40a9ca3b1 service nova] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.468500] env[61243]: DEBUG nova.compute.utils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 712.469885] env[61243]: DEBUG nova.compute.manager [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 712.470072] env[61243]: DEBUG nova.network.neutron [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 712.507284] env[61243]: DEBUG nova.policy [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aa14e81d86db4e6888bd3d3bf8e74417', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b47f117d57e4c82bb34c65843f25ff2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 712.692193] env[61243]: INFO nova.compute.manager [-] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Took 1.02 seconds to deallocate network for instance. [ 712.695662] env[61243]: DEBUG nova.compute.claims [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 712.695922] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.762424] env[61243]: DEBUG nova.network.neutron [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Successfully created port: f8f5b1a3-8bfd-4424-9c46-7497097c2e2c {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 712.817068] env[61243]: DEBUG nova.scheduler.client.report [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.845295] env[61243]: DEBUG oslo_concurrency.lockutils [req-6bad3402-a351-491b-92f8-0eeac98070d6 req-86b7b5ae-923a-44ec-b5e3-96f40a9ca3b1 service nova] Releasing lock "refresh_cache-755cdc16-0eae-4978-917f-634a8ac7af8f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.845545] env[61243]: DEBUG nova.compute.manager [req-6bad3402-a351-491b-92f8-0eeac98070d6 req-86b7b5ae-923a-44ec-b5e3-96f40a9ca3b1 service nova] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Received event network-vif-deleted-304f43ea-8b5b-46a6-9a7c-a355d8f4e546 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 712.973169] env[61243]: DEBUG nova.compute.manager [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 713.329731] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.376s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.329731] env[61243]: DEBUG nova.compute.manager [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 713.330571] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.597s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.332175] env[61243]: INFO nova.compute.claims [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.585667] env[61243]: ERROR nova.compute.manager [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f8f5b1a3-8bfd-4424-9c46-7497097c2e2c, please check neutron logs for more information. [ 713.585667] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 713.585667] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.585667] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 713.585667] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.585667] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 713.585667] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.585667] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 713.585667] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.585667] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 713.585667] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.585667] env[61243]: ERROR nova.compute.manager raise self.value [ 713.585667] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.585667] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 713.585667] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.585667] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 713.586188] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.586188] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 713.586188] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f8f5b1a3-8bfd-4424-9c46-7497097c2e2c, please check neutron logs for more information. [ 713.586188] env[61243]: ERROR nova.compute.manager [ 713.586188] env[61243]: Traceback (most recent call last): [ 713.586188] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 713.586188] env[61243]: listener.cb(fileno) [ 713.586188] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.586188] env[61243]: result = function(*args, **kwargs) [ 713.586188] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 713.586188] env[61243]: return func(*args, **kwargs) [ 713.586188] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 713.586188] env[61243]: raise e [ 713.586188] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.586188] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 713.586188] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.586188] env[61243]: created_port_ids = self._update_ports_for_instance( [ 713.586188] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.586188] env[61243]: with excutils.save_and_reraise_exception(): [ 713.586188] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.586188] env[61243]: self.force_reraise() [ 713.586188] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.586188] env[61243]: raise self.value [ 713.586188] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.586188] env[61243]: updated_port = self._update_port( [ 713.586188] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.586188] env[61243]: _ensure_no_port_binding_failure(port) [ 713.586188] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.586188] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 713.587028] env[61243]: nova.exception.PortBindingFailed: Binding failed for port f8f5b1a3-8bfd-4424-9c46-7497097c2e2c, please check neutron logs for more information. [ 713.587028] env[61243]: Removing descriptor: 16 [ 713.774609] env[61243]: DEBUG nova.compute.manager [req-eb99e55f-ce96-424e-a825-165b0e3da918 req-601ca974-3680-4c89-95d1-1c684aceea26 service nova] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Received event network-changed-f8f5b1a3-8bfd-4424-9c46-7497097c2e2c {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 713.774792] env[61243]: DEBUG nova.compute.manager [req-eb99e55f-ce96-424e-a825-165b0e3da918 req-601ca974-3680-4c89-95d1-1c684aceea26 service nova] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Refreshing instance network info cache due to event network-changed-f8f5b1a3-8bfd-4424-9c46-7497097c2e2c. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 713.775441] env[61243]: DEBUG oslo_concurrency.lockutils [req-eb99e55f-ce96-424e-a825-165b0e3da918 req-601ca974-3680-4c89-95d1-1c684aceea26 service nova] Acquiring lock "refresh_cache-37b43ee9-f991-4d56-8fcc-192f6534f803" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.775598] env[61243]: DEBUG oslo_concurrency.lockutils [req-eb99e55f-ce96-424e-a825-165b0e3da918 req-601ca974-3680-4c89-95d1-1c684aceea26 service nova] Acquired lock "refresh_cache-37b43ee9-f991-4d56-8fcc-192f6534f803" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.775855] env[61243]: DEBUG nova.network.neutron [req-eb99e55f-ce96-424e-a825-165b0e3da918 req-601ca974-3680-4c89-95d1-1c684aceea26 service nova] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Refreshing network info cache for port f8f5b1a3-8bfd-4424-9c46-7497097c2e2c {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 713.838661] env[61243]: DEBUG nova.compute.utils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 713.841845] env[61243]: DEBUG nova.compute.manager [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 713.841938] env[61243]: DEBUG nova.network.neutron [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 713.890018] env[61243]: DEBUG nova.policy [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f541eb1427dc4b79bcbde023ea185130', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ee110d480fff4ca49002ed35f5a0bb47', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 713.983472] env[61243]: DEBUG nova.compute.manager [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 714.012248] env[61243]: DEBUG nova.virt.hardware [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 714.012523] env[61243]: DEBUG nova.virt.hardware [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 714.012718] env[61243]: DEBUG nova.virt.hardware [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 714.013101] env[61243]: DEBUG nova.virt.hardware [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 714.013423] env[61243]: DEBUG nova.virt.hardware [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 714.013499] env[61243]: DEBUG nova.virt.hardware [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 714.013669] env[61243]: DEBUG nova.virt.hardware [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 714.013856] env[61243]: DEBUG nova.virt.hardware [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 714.014042] env[61243]: DEBUG nova.virt.hardware [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 714.014209] env[61243]: DEBUG nova.virt.hardware [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 714.014451] env[61243]: DEBUG nova.virt.hardware [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 714.015949] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464958b7-a97b-461a-a755-c027d0c57ee3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.025501] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3574bc3-caef-45d9-8dea-809e33acc08b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.038964] env[61243]: ERROR nova.compute.manager [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f8f5b1a3-8bfd-4424-9c46-7497097c2e2c, please check neutron logs for more information. [ 714.038964] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Traceback (most recent call last): [ 714.038964] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 714.038964] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] yield resources [ 714.038964] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.038964] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] self.driver.spawn(context, instance, image_meta, [ 714.038964] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 714.038964] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.038964] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.038964] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] vm_ref = self.build_virtual_machine(instance, [ 714.038964] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.039424] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.039424] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.039424] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] for vif in network_info: [ 714.039424] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.039424] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] return self._sync_wrapper(fn, *args, **kwargs) [ 714.039424] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.039424] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] self.wait() [ 714.039424] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.039424] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] self[:] = self._gt.wait() [ 714.039424] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.039424] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] return self._exit_event.wait() [ 714.039424] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 714.039424] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] current.throw(*self._exc) [ 714.039827] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.039827] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] result = function(*args, **kwargs) [ 714.039827] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 714.039827] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] return func(*args, **kwargs) [ 714.039827] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.039827] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] raise e [ 714.039827] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.039827] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] nwinfo = self.network_api.allocate_for_instance( [ 714.039827] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.039827] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] created_port_ids = self._update_ports_for_instance( [ 714.039827] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.039827] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] with excutils.save_and_reraise_exception(): [ 714.039827] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.040315] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] self.force_reraise() [ 714.040315] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.040315] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] raise self.value [ 714.040315] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.040315] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] updated_port = self._update_port( [ 714.040315] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.040315] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] _ensure_no_port_binding_failure(port) [ 714.040315] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.040315] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] raise exception.PortBindingFailed(port_id=port['id']) [ 714.040315] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] nova.exception.PortBindingFailed: Binding failed for port f8f5b1a3-8bfd-4424-9c46-7497097c2e2c, please check neutron logs for more information. [ 714.040315] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] [ 714.040315] env[61243]: INFO nova.compute.manager [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Terminating instance [ 714.041749] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Acquiring lock "refresh_cache-37b43ee9-f991-4d56-8fcc-192f6534f803" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.143857] env[61243]: DEBUG nova.network.neutron [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Successfully created port: 4e9d0e19-cbce-4191-a741-c9a8b8632071 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 714.299105] env[61243]: DEBUG nova.network.neutron [req-eb99e55f-ce96-424e-a825-165b0e3da918 req-601ca974-3680-4c89-95d1-1c684aceea26 service nova] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.342656] env[61243]: DEBUG nova.compute.manager [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 714.428649] env[61243]: DEBUG nova.network.neutron [req-eb99e55f-ce96-424e-a825-165b0e3da918 req-601ca974-3680-4c89-95d1-1c684aceea26 service nova] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.673415] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6fb5b4-c607-4159-87a4-e6c0af45796b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.680993] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee614865-d502-48b5-94ca-97244ad78f6d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.714796] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f94b4b7-a3f3-4237-bb53-e8396cd77130 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.722265] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e46aa87-893c-48a9-a4e4-5cbf80f65675 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.739763] env[61243]: DEBUG nova.compute.provider_tree [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.933874] env[61243]: DEBUG oslo_concurrency.lockutils [req-eb99e55f-ce96-424e-a825-165b0e3da918 req-601ca974-3680-4c89-95d1-1c684aceea26 service nova] Releasing lock "refresh_cache-37b43ee9-f991-4d56-8fcc-192f6534f803" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.934020] env[61243]: DEBUG nova.compute.manager [req-eb99e55f-ce96-424e-a825-165b0e3da918 req-601ca974-3680-4c89-95d1-1c684aceea26 service nova] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Received event network-vif-deleted-f8f5b1a3-8bfd-4424-9c46-7497097c2e2c {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 714.934377] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Acquired lock "refresh_cache-37b43ee9-f991-4d56-8fcc-192f6534f803" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.934725] env[61243]: DEBUG nova.network.neutron [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 715.064308] env[61243]: ERROR nova.compute.manager [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4e9d0e19-cbce-4191-a741-c9a8b8632071, please check neutron logs for more information. [ 715.064308] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 715.064308] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.064308] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 715.064308] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 715.064308] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 715.064308] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 715.064308] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 715.064308] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.064308] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 715.064308] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.064308] env[61243]: ERROR nova.compute.manager raise self.value [ 715.064308] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 715.064308] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 715.064308] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.064308] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 715.064873] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.064873] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 715.064873] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4e9d0e19-cbce-4191-a741-c9a8b8632071, please check neutron logs for more information. [ 715.064873] env[61243]: ERROR nova.compute.manager [ 715.064873] env[61243]: Traceback (most recent call last): [ 715.064873] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 715.064873] env[61243]: listener.cb(fileno) [ 715.064873] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 715.064873] env[61243]: result = function(*args, **kwargs) [ 715.064873] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 715.064873] env[61243]: return func(*args, **kwargs) [ 715.064873] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 715.064873] env[61243]: raise e [ 715.064873] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.064873] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 715.064873] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 715.064873] env[61243]: created_port_ids = self._update_ports_for_instance( [ 715.064873] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 715.064873] env[61243]: with excutils.save_and_reraise_exception(): [ 715.064873] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.064873] env[61243]: self.force_reraise() [ 715.064873] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.064873] env[61243]: raise self.value [ 715.064873] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 715.064873] env[61243]: updated_port = self._update_port( [ 715.064873] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.064873] env[61243]: _ensure_no_port_binding_failure(port) [ 715.064873] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.064873] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 715.065813] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 4e9d0e19-cbce-4191-a741-c9a8b8632071, please check neutron logs for more information. [ 715.065813] env[61243]: Removing descriptor: 16 [ 715.244053] env[61243]: DEBUG nova.scheduler.client.report [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 715.353989] env[61243]: DEBUG nova.compute.manager [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 715.380235] env[61243]: DEBUG nova.virt.hardware [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 715.380480] env[61243]: DEBUG nova.virt.hardware [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 715.380636] env[61243]: DEBUG nova.virt.hardware [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 715.380818] env[61243]: DEBUG nova.virt.hardware [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 715.380969] env[61243]: DEBUG nova.virt.hardware [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 715.381170] env[61243]: DEBUG nova.virt.hardware [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 715.381382] env[61243]: DEBUG nova.virt.hardware [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 715.381539] env[61243]: DEBUG nova.virt.hardware [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 715.381734] env[61243]: DEBUG nova.virt.hardware [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 715.381903] env[61243]: DEBUG nova.virt.hardware [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 715.382092] env[61243]: DEBUG nova.virt.hardware [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 715.383017] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f06d4c93-909d-4eaf-901d-646e8b1f1068 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.390944] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72def81-fa3a-4122-84b4-0aafa286d86a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.404670] env[61243]: ERROR nova.compute.manager [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4e9d0e19-cbce-4191-a741-c9a8b8632071, please check neutron logs for more information. [ 715.404670] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Traceback (most recent call last): [ 715.404670] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 715.404670] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] yield resources [ 715.404670] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 715.404670] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] self.driver.spawn(context, instance, image_meta, [ 715.404670] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 715.404670] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 715.404670] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 715.404670] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] vm_ref = self.build_virtual_machine(instance, [ 715.404670] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 715.405093] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 715.405093] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 715.405093] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] for vif in network_info: [ 715.405093] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 715.405093] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] return self._sync_wrapper(fn, *args, **kwargs) [ 715.405093] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 715.405093] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] self.wait() [ 715.405093] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 715.405093] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] self[:] = self._gt.wait() [ 715.405093] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 715.405093] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] return self._exit_event.wait() [ 715.405093] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 715.405093] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] current.throw(*self._exc) [ 715.405496] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 715.405496] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] result = function(*args, **kwargs) [ 715.405496] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 715.405496] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] return func(*args, **kwargs) [ 715.405496] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 715.405496] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] raise e [ 715.405496] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.405496] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] nwinfo = self.network_api.allocate_for_instance( [ 715.405496] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 715.405496] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] created_port_ids = self._update_ports_for_instance( [ 715.405496] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 715.405496] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] with excutils.save_and_reraise_exception(): [ 715.405496] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.405903] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] self.force_reraise() [ 715.405903] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.405903] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] raise self.value [ 715.405903] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 715.405903] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] updated_port = self._update_port( [ 715.405903] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.405903] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] _ensure_no_port_binding_failure(port) [ 715.405903] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.405903] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] raise exception.PortBindingFailed(port_id=port['id']) [ 715.405903] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] nova.exception.PortBindingFailed: Binding failed for port 4e9d0e19-cbce-4191-a741-c9a8b8632071, please check neutron logs for more information. [ 715.405903] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] [ 715.405903] env[61243]: INFO nova.compute.manager [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Terminating instance [ 715.406791] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Acquiring lock "refresh_cache-8e85f285-1190-4c15-bc1a-8364434dc4ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.406964] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Acquired lock "refresh_cache-8e85f285-1190-4c15-bc1a-8364434dc4ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.407151] env[61243]: DEBUG nova.network.neutron [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 715.452374] env[61243]: DEBUG nova.network.neutron [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.526248] env[61243]: DEBUG nova.network.neutron [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.748518] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.749038] env[61243]: DEBUG nova.compute.manager [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 715.751504] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.874s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.752916] env[61243]: INFO nova.compute.claims [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 715.799728] env[61243]: DEBUG nova.compute.manager [req-994992c1-7eb7-44da-82fc-50acee7f035b req-f98b217b-648f-4a88-b3be-666f689ffc96 service nova] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Received event network-changed-4e9d0e19-cbce-4191-a741-c9a8b8632071 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 715.799973] env[61243]: DEBUG nova.compute.manager [req-994992c1-7eb7-44da-82fc-50acee7f035b req-f98b217b-648f-4a88-b3be-666f689ffc96 service nova] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Refreshing instance network info cache due to event network-changed-4e9d0e19-cbce-4191-a741-c9a8b8632071. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 715.800165] env[61243]: DEBUG oslo_concurrency.lockutils [req-994992c1-7eb7-44da-82fc-50acee7f035b req-f98b217b-648f-4a88-b3be-666f689ffc96 service nova] Acquiring lock "refresh_cache-8e85f285-1190-4c15-bc1a-8364434dc4ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.924706] env[61243]: DEBUG nova.network.neutron [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.008986] env[61243]: DEBUG nova.network.neutron [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.028829] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Releasing lock "refresh_cache-37b43ee9-f991-4d56-8fcc-192f6534f803" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.029270] env[61243]: DEBUG nova.compute.manager [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 716.029454] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 716.029749] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-36bdeff3-a061-4560-8a34-688d3c2476e4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.038770] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd0109c-0788-4b83-a099-1a426b9ed7b2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.067622] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 37b43ee9-f991-4d56-8fcc-192f6534f803 could not be found. [ 716.068202] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 716.068202] env[61243]: INFO nova.compute.manager [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Took 0.04 seconds to destroy the instance on the hypervisor. [ 716.068382] env[61243]: DEBUG oslo.service.loopingcall [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.068517] env[61243]: DEBUG nova.compute.manager [-] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 716.068610] env[61243]: DEBUG nova.network.neutron [-] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 716.083543] env[61243]: DEBUG nova.network.neutron [-] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.257556] env[61243]: DEBUG nova.compute.utils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 716.261028] env[61243]: DEBUG nova.compute.manager [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 716.261028] env[61243]: DEBUG nova.network.neutron [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 716.299714] env[61243]: DEBUG nova.policy [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f541eb1427dc4b79bcbde023ea185130', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ee110d480fff4ca49002ed35f5a0bb47', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 716.511754] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Releasing lock "refresh_cache-8e85f285-1190-4c15-bc1a-8364434dc4ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.512388] env[61243]: DEBUG nova.compute.manager [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 716.512477] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 716.512754] env[61243]: DEBUG oslo_concurrency.lockutils [req-994992c1-7eb7-44da-82fc-50acee7f035b req-f98b217b-648f-4a88-b3be-666f689ffc96 service nova] Acquired lock "refresh_cache-8e85f285-1190-4c15-bc1a-8364434dc4ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.512933] env[61243]: DEBUG nova.network.neutron [req-994992c1-7eb7-44da-82fc-50acee7f035b req-f98b217b-648f-4a88-b3be-666f689ffc96 service nova] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Refreshing network info cache for port 4e9d0e19-cbce-4191-a741-c9a8b8632071 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 716.514050] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c2885dc9-5ff6-4f8e-9114-124fb4e35269 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.522675] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499bd209-ce45-4e24-83db-6d3212ad769e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.545044] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8e85f285-1190-4c15-bc1a-8364434dc4ba could not be found. [ 716.545283] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 716.545485] env[61243]: INFO nova.compute.manager [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Took 0.03 seconds to destroy the instance on the hypervisor. [ 716.545734] env[61243]: DEBUG oslo.service.loopingcall [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.545998] env[61243]: DEBUG nova.compute.manager [-] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 716.546114] env[61243]: DEBUG nova.network.neutron [-] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 716.551021] env[61243]: DEBUG nova.network.neutron [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Successfully created port: 779de90f-7a8d-43b5-a1ba-b24777560cfd {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 716.570729] env[61243]: DEBUG nova.network.neutron [-] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.585619] env[61243]: DEBUG nova.network.neutron [-] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.762053] env[61243]: DEBUG nova.compute.manager [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 717.037483] env[61243]: DEBUG nova.network.neutron [req-994992c1-7eb7-44da-82fc-50acee7f035b req-f98b217b-648f-4a88-b3be-666f689ffc96 service nova] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.075465] env[61243]: DEBUG nova.network.neutron [-] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.083305] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da3a03ad-803e-45b3-a654-8392dfdfd704 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.088220] env[61243]: INFO nova.compute.manager [-] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Took 1.02 seconds to deallocate network for instance. [ 717.094092] env[61243]: DEBUG nova.compute.claims [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 717.094584] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.099090] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d407f3-3a53-4e6e-922c-61a401cc6dbf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.132046] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a6541e1-5a32-46cf-8c03-adac333adc24 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.139260] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a4502d-7b76-4764-928c-13ea98604d33 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.155080] env[61243]: DEBUG nova.compute.provider_tree [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.156992] env[61243]: DEBUG nova.network.neutron [req-994992c1-7eb7-44da-82fc-50acee7f035b req-f98b217b-648f-4a88-b3be-666f689ffc96 service nova] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.462502] env[61243]: ERROR nova.compute.manager [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 779de90f-7a8d-43b5-a1ba-b24777560cfd, please check neutron logs for more information. [ 717.462502] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 717.462502] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.462502] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 717.462502] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.462502] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 717.462502] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.462502] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 717.462502] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.462502] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 717.462502] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.462502] env[61243]: ERROR nova.compute.manager raise self.value [ 717.462502] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.462502] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 717.462502] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.462502] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 717.463035] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.463035] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 717.463035] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 779de90f-7a8d-43b5-a1ba-b24777560cfd, please check neutron logs for more information. [ 717.463035] env[61243]: ERROR nova.compute.manager [ 717.463035] env[61243]: Traceback (most recent call last): [ 717.463035] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 717.463035] env[61243]: listener.cb(fileno) [ 717.463035] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.463035] env[61243]: result = function(*args, **kwargs) [ 717.463035] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 717.463035] env[61243]: return func(*args, **kwargs) [ 717.463035] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.463035] env[61243]: raise e [ 717.463035] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.463035] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 717.463035] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.463035] env[61243]: created_port_ids = self._update_ports_for_instance( [ 717.463035] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.463035] env[61243]: with excutils.save_and_reraise_exception(): [ 717.463035] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.463035] env[61243]: self.force_reraise() [ 717.463035] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.463035] env[61243]: raise self.value [ 717.463035] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.463035] env[61243]: updated_port = self._update_port( [ 717.463035] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.463035] env[61243]: _ensure_no_port_binding_failure(port) [ 717.463035] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.463035] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 717.463898] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 779de90f-7a8d-43b5-a1ba-b24777560cfd, please check neutron logs for more information. [ 717.463898] env[61243]: Removing descriptor: 16 [ 717.577975] env[61243]: INFO nova.compute.manager [-] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Took 1.03 seconds to deallocate network for instance. [ 717.580334] env[61243]: DEBUG nova.compute.claims [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 717.580519] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.660180] env[61243]: DEBUG nova.scheduler.client.report [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 717.664057] env[61243]: DEBUG oslo_concurrency.lockutils [req-994992c1-7eb7-44da-82fc-50acee7f035b req-f98b217b-648f-4a88-b3be-666f689ffc96 service nova] Releasing lock "refresh_cache-8e85f285-1190-4c15-bc1a-8364434dc4ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.664057] env[61243]: DEBUG nova.compute.manager [req-994992c1-7eb7-44da-82fc-50acee7f035b req-f98b217b-648f-4a88-b3be-666f689ffc96 service nova] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Received event network-vif-deleted-4e9d0e19-cbce-4191-a741-c9a8b8632071 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 717.773217] env[61243]: DEBUG nova.compute.manager [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 717.797814] env[61243]: DEBUG nova.virt.hardware [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 717.798072] env[61243]: DEBUG nova.virt.hardware [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 717.798233] env[61243]: DEBUG nova.virt.hardware [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 717.798416] env[61243]: DEBUG nova.virt.hardware [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 717.798563] env[61243]: DEBUG nova.virt.hardware [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 717.798713] env[61243]: DEBUG nova.virt.hardware [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 717.798922] env[61243]: DEBUG nova.virt.hardware [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 717.799091] env[61243]: DEBUG nova.virt.hardware [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 717.799259] env[61243]: DEBUG nova.virt.hardware [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 717.799420] env[61243]: DEBUG nova.virt.hardware [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 717.799611] env[61243]: DEBUG nova.virt.hardware [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 717.800500] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f8b7cb-4daa-4ee6-bed4-1a791cbeb116 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.808704] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9264a508-3b4c-4b73-b5be-43073cd53d21 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.823090] env[61243]: ERROR nova.compute.manager [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 779de90f-7a8d-43b5-a1ba-b24777560cfd, please check neutron logs for more information. [ 717.823090] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Traceback (most recent call last): [ 717.823090] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 717.823090] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] yield resources [ 717.823090] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 717.823090] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] self.driver.spawn(context, instance, image_meta, [ 717.823090] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 717.823090] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.823090] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.823090] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] vm_ref = self.build_virtual_machine(instance, [ 717.823090] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.823487] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.823487] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.823487] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] for vif in network_info: [ 717.823487] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 717.823487] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] return self._sync_wrapper(fn, *args, **kwargs) [ 717.823487] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 717.823487] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] self.wait() [ 717.823487] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 717.823487] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] self[:] = self._gt.wait() [ 717.823487] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.823487] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] return self._exit_event.wait() [ 717.823487] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 717.823487] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] current.throw(*self._exc) [ 717.824148] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.824148] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] result = function(*args, **kwargs) [ 717.824148] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 717.824148] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] return func(*args, **kwargs) [ 717.824148] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.824148] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] raise e [ 717.824148] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.824148] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] nwinfo = self.network_api.allocate_for_instance( [ 717.824148] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.824148] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] created_port_ids = self._update_ports_for_instance( [ 717.824148] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.824148] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] with excutils.save_and_reraise_exception(): [ 717.824148] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.824618] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] self.force_reraise() [ 717.824618] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.824618] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] raise self.value [ 717.824618] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.824618] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] updated_port = self._update_port( [ 717.824618] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.824618] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] _ensure_no_port_binding_failure(port) [ 717.824618] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.824618] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] raise exception.PortBindingFailed(port_id=port['id']) [ 717.824618] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] nova.exception.PortBindingFailed: Binding failed for port 779de90f-7a8d-43b5-a1ba-b24777560cfd, please check neutron logs for more information. [ 717.824618] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] [ 717.824618] env[61243]: INFO nova.compute.manager [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Terminating instance [ 717.826308] env[61243]: DEBUG nova.compute.manager [req-7d193b14-e885-4741-bab4-cd7ff3a501c5 req-aa6f7b9e-0c28-4dc2-9bb7-304818e86972 service nova] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Received event network-changed-779de90f-7a8d-43b5-a1ba-b24777560cfd {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 717.826490] env[61243]: DEBUG nova.compute.manager [req-7d193b14-e885-4741-bab4-cd7ff3a501c5 req-aa6f7b9e-0c28-4dc2-9bb7-304818e86972 service nova] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Refreshing instance network info cache due to event network-changed-779de90f-7a8d-43b5-a1ba-b24777560cfd. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 717.826696] env[61243]: DEBUG oslo_concurrency.lockutils [req-7d193b14-e885-4741-bab4-cd7ff3a501c5 req-aa6f7b9e-0c28-4dc2-9bb7-304818e86972 service nova] Acquiring lock "refresh_cache-f6566092-9921-4a94-a0b5-a2d3803a9d18" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.826840] env[61243]: DEBUG oslo_concurrency.lockutils [req-7d193b14-e885-4741-bab4-cd7ff3a501c5 req-aa6f7b9e-0c28-4dc2-9bb7-304818e86972 service nova] Acquired lock "refresh_cache-f6566092-9921-4a94-a0b5-a2d3803a9d18" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.826994] env[61243]: DEBUG nova.network.neutron [req-7d193b14-e885-4741-bab4-cd7ff3a501c5 req-aa6f7b9e-0c28-4dc2-9bb7-304818e86972 service nova] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Refreshing network info cache for port 779de90f-7a8d-43b5-a1ba-b24777560cfd {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 717.828048] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Acquiring lock "refresh_cache-f6566092-9921-4a94-a0b5-a2d3803a9d18" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.166033] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.414s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.166142] env[61243]: DEBUG nova.compute.manager [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 718.168747] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.901s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.346503] env[61243]: DEBUG nova.network.neutron [req-7d193b14-e885-4741-bab4-cd7ff3a501c5 req-aa6f7b9e-0c28-4dc2-9bb7-304818e86972 service nova] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.426193] env[61243]: DEBUG nova.network.neutron [req-7d193b14-e885-4741-bab4-cd7ff3a501c5 req-aa6f7b9e-0c28-4dc2-9bb7-304818e86972 service nova] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.673225] env[61243]: DEBUG nova.compute.utils [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 718.675136] env[61243]: DEBUG nova.compute.manager [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Not allocating networking since 'none' was specified. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 718.931312] env[61243]: DEBUG oslo_concurrency.lockutils [req-7d193b14-e885-4741-bab4-cd7ff3a501c5 req-aa6f7b9e-0c28-4dc2-9bb7-304818e86972 service nova] Releasing lock "refresh_cache-f6566092-9921-4a94-a0b5-a2d3803a9d18" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.931312] env[61243]: DEBUG nova.compute.manager [req-7d193b14-e885-4741-bab4-cd7ff3a501c5 req-aa6f7b9e-0c28-4dc2-9bb7-304818e86972 service nova] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Received event network-vif-deleted-779de90f-7a8d-43b5-a1ba-b24777560cfd {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.931312] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Acquired lock "refresh_cache-f6566092-9921-4a94-a0b5-a2d3803a9d18" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.931484] env[61243]: DEBUG nova.network.neutron [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 718.992482] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c805f2f4-e006-4f8d-ab27-9f6eb81ff3d4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.000640] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-020ecd97-45ed-4195-9038-61816b4f8607 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.031041] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ac1b90-8b3d-4409-a76f-a79f57f9ebc3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.038123] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e71e94e-d62e-4689-8a68-0d2fd1cb2853 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.052172] env[61243]: DEBUG nova.compute.provider_tree [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.179224] env[61243]: DEBUG nova.compute.manager [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 719.449565] env[61243]: DEBUG nova.network.neutron [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.526556] env[61243]: DEBUG nova.network.neutron [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.560710] env[61243]: DEBUG nova.scheduler.client.report [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.030276] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Releasing lock "refresh_cache-f6566092-9921-4a94-a0b5-a2d3803a9d18" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.030673] env[61243]: DEBUG nova.compute.manager [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 720.030871] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 720.031194] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8719e93c-9e22-49fa-9bad-2d2fb3d62c9a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.040137] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2eb5aa7-d816-4e37-9691-5c12d70f23f8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.060708] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f6566092-9921-4a94-a0b5-a2d3803a9d18 could not be found. [ 720.060918] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 720.061118] env[61243]: INFO nova.compute.manager [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Took 0.03 seconds to destroy the instance on the hypervisor. [ 720.061359] env[61243]: DEBUG oslo.service.loopingcall [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.061575] env[61243]: DEBUG nova.compute.manager [-] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.061669] env[61243]: DEBUG nova.network.neutron [-] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 720.066080] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.897s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.066629] env[61243]: ERROR nova.compute.manager [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9d5a7a9b-a4c9-4217-b740-0c33d86109a2, please check neutron logs for more information. [ 720.066629] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Traceback (most recent call last): [ 720.066629] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 720.066629] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] self.driver.spawn(context, instance, image_meta, [ 720.066629] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 720.066629] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.066629] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.066629] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] vm_ref = self.build_virtual_machine(instance, [ 720.066629] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.066629] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.066629] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.066960] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] for vif in network_info: [ 720.066960] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.066960] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] return self._sync_wrapper(fn, *args, **kwargs) [ 720.066960] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.066960] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] self.wait() [ 720.066960] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.066960] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] self[:] = self._gt.wait() [ 720.066960] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.066960] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] return self._exit_event.wait() [ 720.066960] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 720.066960] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] current.throw(*self._exc) [ 720.066960] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.066960] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] result = function(*args, **kwargs) [ 720.067581] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 720.067581] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] return func(*args, **kwargs) [ 720.067581] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.067581] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] raise e [ 720.067581] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.067581] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] nwinfo = self.network_api.allocate_for_instance( [ 720.067581] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.067581] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] created_port_ids = self._update_ports_for_instance( [ 720.067581] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.067581] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] with excutils.save_and_reraise_exception(): [ 720.067581] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.067581] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] self.force_reraise() [ 720.067581] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.068142] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] raise self.value [ 720.068142] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.068142] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] updated_port = self._update_port( [ 720.068142] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.068142] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] _ensure_no_port_binding_failure(port) [ 720.068142] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.068142] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] raise exception.PortBindingFailed(port_id=port['id']) [ 720.068142] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] nova.exception.PortBindingFailed: Binding failed for port 9d5a7a9b-a4c9-4217-b740-0c33d86109a2, please check neutron logs for more information. [ 720.068142] env[61243]: ERROR nova.compute.manager [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] [ 720.068142] env[61243]: DEBUG nova.compute.utils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Binding failed for port 9d5a7a9b-a4c9-4217-b740-0c33d86109a2, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 720.068579] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.014s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.071186] env[61243]: INFO nova.compute.claims [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 720.073958] env[61243]: DEBUG nova.compute.manager [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Build of instance 72aa2816-5e02-4dc2-9d6a-c3e58c651224 was re-scheduled: Binding failed for port 9d5a7a9b-a4c9-4217-b740-0c33d86109a2, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 720.074417] env[61243]: DEBUG nova.compute.manager [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 720.074600] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Acquiring lock "refresh_cache-72aa2816-5e02-4dc2-9d6a-c3e58c651224" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.074749] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Acquired lock "refresh_cache-72aa2816-5e02-4dc2-9d6a-c3e58c651224" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.074908] env[61243]: DEBUG nova.network.neutron [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 720.076974] env[61243]: DEBUG nova.network.neutron [-] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.193699] env[61243]: DEBUG nova.compute.manager [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 720.218203] env[61243]: DEBUG nova.virt.hardware [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 720.218485] env[61243]: DEBUG nova.virt.hardware [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 720.218650] env[61243]: DEBUG nova.virt.hardware [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.218957] env[61243]: DEBUG nova.virt.hardware [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 720.219176] env[61243]: DEBUG nova.virt.hardware [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.219337] env[61243]: DEBUG nova.virt.hardware [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 720.219553] env[61243]: DEBUG nova.virt.hardware [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 720.219716] env[61243]: DEBUG nova.virt.hardware [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 720.219883] env[61243]: DEBUG nova.virt.hardware [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 720.220058] env[61243]: DEBUG nova.virt.hardware [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 720.220237] env[61243]: DEBUG nova.virt.hardware [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.221101] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f9a87fd-2cc4-4fc7-ae54-e2b58f12b801 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.229188] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64438fcd-e39a-49be-88d9-bddc3af2e440 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.242548] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Instance VIF info [] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 720.247946] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Creating folder: Project (ce559c68ad90452b82dd12f07b0c11a9). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 720.248230] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eefbc507-a53e-43ab-bb9b-d16e70d2a0e5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.261572] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Created folder: Project (ce559c68ad90452b82dd12f07b0c11a9) in parent group-v285636. [ 720.261826] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Creating folder: Instances. Parent ref: group-v285653. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 720.262062] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2475074b-5296-4e16-bbf9-94a97ebf5438 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.271060] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Created folder: Instances in parent group-v285653. [ 720.271294] env[61243]: DEBUG oslo.service.loopingcall [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.271480] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 720.271673] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-af315271-7ae1-482e-ac63-76a4f59ffd98 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.289016] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 720.289016] env[61243]: value = "task-1338750" [ 720.289016] env[61243]: _type = "Task" [ 720.289016] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.296082] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338750, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.581758] env[61243]: DEBUG nova.network.neutron [-] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.594345] env[61243]: DEBUG nova.network.neutron [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.661116] env[61243]: DEBUG nova.network.neutron [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.798606] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338750, 'name': CreateVM_Task, 'duration_secs': 0.267527} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.798804] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 720.799225] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.799385] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.799738] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 720.800293] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-804b0601-c9b0-4a36-99b2-9c7a342d2e1d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.804455] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Waiting for the task: (returnval){ [ 720.804455] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5250d2d2-1ee5-9f1b-7c71-35da87e96dd4" [ 720.804455] env[61243]: _type = "Task" [ 720.804455] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.811557] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5250d2d2-1ee5-9f1b-7c71-35da87e96dd4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.084014] env[61243]: INFO nova.compute.manager [-] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Took 1.02 seconds to deallocate network for instance. [ 721.086238] env[61243]: DEBUG nova.compute.claims [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 721.086416] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.163719] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Releasing lock "refresh_cache-72aa2816-5e02-4dc2-9d6a-c3e58c651224" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.163719] env[61243]: DEBUG nova.compute.manager [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 721.163719] env[61243]: DEBUG nova.compute.manager [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 721.163960] env[61243]: DEBUG nova.network.neutron [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 721.183664] env[61243]: DEBUG nova.network.neutron [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.316444] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5250d2d2-1ee5-9f1b-7c71-35da87e96dd4, 'name': SearchDatastore_Task, 'duration_secs': 0.012204} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.316744] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.316979] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 721.317222] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.317368] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.317540] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 721.317789] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eb797097-7dda-4116-9079-94e35f9578e7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.328068] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 721.328308] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 721.329046] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c19c0258-c2e2-46b1-8baf-ee9f8359cec1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.336494] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Waiting for the task: (returnval){ [ 721.336494] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52db1451-4f56-4bd7-d091-e3118de20dd2" [ 721.336494] env[61243]: _type = "Task" [ 721.336494] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.346271] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52db1451-4f56-4bd7-d091-e3118de20dd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.373923] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050999df-a526-4303-bd64-9a6e6cb1f80d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.379957] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac8307ec-4d41-4cac-8c02-4dda644f28d2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.410249] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d181015d-6ccb-400c-b763-d83ab85fe7f8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.417670] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd70e047-1330-4884-a346-2b0534f6dae0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.430357] env[61243]: DEBUG nova.compute.provider_tree [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.687019] env[61243]: DEBUG nova.network.neutron [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.847025] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52db1451-4f56-4bd7-d091-e3118de20dd2, 'name': SearchDatastore_Task, 'duration_secs': 0.009356} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.847808] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-338a7072-e627-414d-a4b8-ec15f2900681 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.852734] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Waiting for the task: (returnval){ [ 721.852734] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52acd973-8bab-dd12-f058-7b0c37f0bd32" [ 721.852734] env[61243]: _type = "Task" [ 721.852734] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.859834] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52acd973-8bab-dd12-f058-7b0c37f0bd32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.933404] env[61243]: DEBUG nova.scheduler.client.report [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 722.188675] env[61243]: INFO nova.compute.manager [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] [instance: 72aa2816-5e02-4dc2-9d6a-c3e58c651224] Took 1.02 seconds to deallocate network for instance. [ 722.363587] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52acd973-8bab-dd12-f058-7b0c37f0bd32, 'name': SearchDatastore_Task, 'duration_secs': 0.011229} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.363844] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.364122] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] fd2d724a-76b6-405f-b003-54de2ad84f5a/fd2d724a-76b6-405f-b003-54de2ad84f5a.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 722.364371] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9ee16f6-d4ef-458d-a585-4baaf82bfb44 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.370985] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Waiting for the task: (returnval){ [ 722.370985] env[61243]: value = "task-1338751" [ 722.370985] env[61243]: _type = "Task" [ 722.370985] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.378407] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': task-1338751, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.438659] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.370s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.438843] env[61243]: DEBUG nova.compute.manager [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 722.441592] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.684s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.881685] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': task-1338751, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497289} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.882770] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] fd2d724a-76b6-405f-b003-54de2ad84f5a/fd2d724a-76b6-405f-b003-54de2ad84f5a.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 722.882770] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 722.882770] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5d6f07fe-1060-4a64-a27a-ff3f2362ff02 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.888475] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Waiting for the task: (returnval){ [ 722.888475] env[61243]: value = "task-1338752" [ 722.888475] env[61243]: _type = "Task" [ 722.888475] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.898039] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': task-1338752, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.946806] env[61243]: DEBUG nova.compute.utils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 722.951927] env[61243]: DEBUG nova.compute.manager [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 722.952129] env[61243]: DEBUG nova.network.neutron [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 723.000455] env[61243]: DEBUG nova.policy [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '80d738b3541149e38c138cc2d6acdb79', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '978e2923cf4141049571930c41db16a9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 723.216085] env[61243]: INFO nova.scheduler.client.report [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Deleted allocations for instance 72aa2816-5e02-4dc2-9d6a-c3e58c651224 [ 723.284046] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-347b2b8a-568a-4202-800e-0a3831a637d5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.292644] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9bc1d4-dccf-4780-a666-f9dd88104564 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.296923] env[61243]: DEBUG nova.network.neutron [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Successfully created port: 08bf97d6-7dce-4603-9977-136dce2b29cb {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 723.327039] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09441420-9620-45a4-87cb-29e7d51a9aba {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.334183] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7f2743-028b-4993-825d-022a3cd9480e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.346999] env[61243]: DEBUG nova.compute.provider_tree [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.397857] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': task-1338752, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063455} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.398193] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 723.398957] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36b8671f-cc3c-4145-b489-45eb6bb95745 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.417652] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] fd2d724a-76b6-405f-b003-54de2ad84f5a/fd2d724a-76b6-405f-b003-54de2ad84f5a.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 723.417888] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f0e8e3c-d484-4eec-8f75-f98c0f1cee5e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.437461] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Waiting for the task: (returnval){ [ 723.437461] env[61243]: value = "task-1338753" [ 723.437461] env[61243]: _type = "Task" [ 723.437461] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.446289] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': task-1338753, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.454995] env[61243]: DEBUG nova.compute.manager [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 723.726419] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f223d882-ce1b-475e-aecf-5c93d2e4cdea tempest-FloatingIPsAssociationNegativeTestJSON-784908522 tempest-FloatingIPsAssociationNegativeTestJSON-784908522-project-member] Lock "72aa2816-5e02-4dc2-9d6a-c3e58c651224" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.337s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.852083] env[61243]: DEBUG nova.scheduler.client.report [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.950721] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': task-1338753, 'name': ReconfigVM_Task, 'duration_secs': 0.262557} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.951029] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Reconfigured VM instance instance-00000027 to attach disk [datastore2] fd2d724a-76b6-405f-b003-54de2ad84f5a/fd2d724a-76b6-405f-b003-54de2ad84f5a.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 723.951659] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d4bffe25-0bca-4dfa-880c-1e5a505b4ea6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.958809] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Waiting for the task: (returnval){ [ 723.958809] env[61243]: value = "task-1338754" [ 723.958809] env[61243]: _type = "Task" [ 723.958809] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.970179] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': task-1338754, 'name': Rename_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.002463] env[61243]: DEBUG nova.compute.manager [req-226568ec-d470-4ac5-8ebe-afdc6b5a3782 req-c856f5a4-35ae-4545-8103-bd424d89af1f service nova] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Received event network-changed-08bf97d6-7dce-4603-9977-136dce2b29cb {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 724.002748] env[61243]: DEBUG nova.compute.manager [req-226568ec-d470-4ac5-8ebe-afdc6b5a3782 req-c856f5a4-35ae-4545-8103-bd424d89af1f service nova] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Refreshing instance network info cache due to event network-changed-08bf97d6-7dce-4603-9977-136dce2b29cb. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 724.003010] env[61243]: DEBUG oslo_concurrency.lockutils [req-226568ec-d470-4ac5-8ebe-afdc6b5a3782 req-c856f5a4-35ae-4545-8103-bd424d89af1f service nova] Acquiring lock "refresh_cache-3b852a19-9f86-486b-b98d-e87ffc54d943" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.003325] env[61243]: DEBUG oslo_concurrency.lockutils [req-226568ec-d470-4ac5-8ebe-afdc6b5a3782 req-c856f5a4-35ae-4545-8103-bd424d89af1f service nova] Acquired lock "refresh_cache-3b852a19-9f86-486b-b98d-e87ffc54d943" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.003325] env[61243]: DEBUG nova.network.neutron [req-226568ec-d470-4ac5-8ebe-afdc6b5a3782 req-c856f5a4-35ae-4545-8103-bd424d89af1f service nova] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Refreshing network info cache for port 08bf97d6-7dce-4603-9977-136dce2b29cb {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 724.135242] env[61243]: ERROR nova.compute.manager [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 08bf97d6-7dce-4603-9977-136dce2b29cb, please check neutron logs for more information. [ 724.135242] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 724.135242] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.135242] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 724.135242] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.135242] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 724.135242] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.135242] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 724.135242] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.135242] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 724.135242] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.135242] env[61243]: ERROR nova.compute.manager raise self.value [ 724.135242] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.135242] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 724.135242] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.135242] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 724.135779] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.135779] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 724.135779] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 08bf97d6-7dce-4603-9977-136dce2b29cb, please check neutron logs for more information. [ 724.135779] env[61243]: ERROR nova.compute.manager [ 724.135779] env[61243]: Traceback (most recent call last): [ 724.135779] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 724.135779] env[61243]: listener.cb(fileno) [ 724.135779] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.135779] env[61243]: result = function(*args, **kwargs) [ 724.135779] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 724.135779] env[61243]: return func(*args, **kwargs) [ 724.135779] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.135779] env[61243]: raise e [ 724.135779] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.135779] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 724.135779] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.135779] env[61243]: created_port_ids = self._update_ports_for_instance( [ 724.135779] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.135779] env[61243]: with excutils.save_and_reraise_exception(): [ 724.135779] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.135779] env[61243]: self.force_reraise() [ 724.135779] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.135779] env[61243]: raise self.value [ 724.135779] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.135779] env[61243]: updated_port = self._update_port( [ 724.135779] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.135779] env[61243]: _ensure_no_port_binding_failure(port) [ 724.135779] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.135779] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 724.136685] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 08bf97d6-7dce-4603-9977-136dce2b29cb, please check neutron logs for more information. [ 724.136685] env[61243]: Removing descriptor: 16 [ 724.230180] env[61243]: DEBUG nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 724.358178] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.916s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.358602] env[61243]: ERROR nova.compute.manager [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bfc0ef48-c4b5-4571-8059-4253feb36455, please check neutron logs for more information. [ 724.358602] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Traceback (most recent call last): [ 724.358602] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.358602] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] self.driver.spawn(context, instance, image_meta, [ 724.358602] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 724.358602] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.358602] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.358602] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] vm_ref = self.build_virtual_machine(instance, [ 724.358602] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.358602] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.358602] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.358944] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] for vif in network_info: [ 724.358944] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.358944] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] return self._sync_wrapper(fn, *args, **kwargs) [ 724.358944] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.358944] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] self.wait() [ 724.358944] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.358944] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] self[:] = self._gt.wait() [ 724.358944] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.358944] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] return self._exit_event.wait() [ 724.358944] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 724.358944] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] current.throw(*self._exc) [ 724.358944] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.358944] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] result = function(*args, **kwargs) [ 724.359362] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 724.359362] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] return func(*args, **kwargs) [ 724.359362] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.359362] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] raise e [ 724.359362] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.359362] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] nwinfo = self.network_api.allocate_for_instance( [ 724.359362] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.359362] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] created_port_ids = self._update_ports_for_instance( [ 724.359362] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.359362] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] with excutils.save_and_reraise_exception(): [ 724.359362] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.359362] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] self.force_reraise() [ 724.359362] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.359697] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] raise self.value [ 724.359697] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.359697] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] updated_port = self._update_port( [ 724.359697] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.359697] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] _ensure_no_port_binding_failure(port) [ 724.359697] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.359697] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] raise exception.PortBindingFailed(port_id=port['id']) [ 724.359697] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] nova.exception.PortBindingFailed: Binding failed for port bfc0ef48-c4b5-4571-8059-4253feb36455, please check neutron logs for more information. [ 724.359697] env[61243]: ERROR nova.compute.manager [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] [ 724.359697] env[61243]: DEBUG nova.compute.utils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Binding failed for port bfc0ef48-c4b5-4571-8059-4253feb36455, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 724.360651] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.976s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.362192] env[61243]: INFO nova.compute.claims [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 724.364731] env[61243]: DEBUG nova.compute.manager [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Build of instance c887c01b-a67e-4e3e-9e17-1deb26e9b508 was re-scheduled: Binding failed for port bfc0ef48-c4b5-4571-8059-4253feb36455, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 724.365172] env[61243]: DEBUG nova.compute.manager [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 724.365407] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Acquiring lock "refresh_cache-c887c01b-a67e-4e3e-9e17-1deb26e9b508" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.365591] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Acquired lock "refresh_cache-c887c01b-a67e-4e3e-9e17-1deb26e9b508" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.365764] env[61243]: DEBUG nova.network.neutron [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 724.471440] env[61243]: DEBUG nova.compute.manager [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 724.471440] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': task-1338754, 'name': Rename_Task, 'duration_secs': 0.1425} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.474170] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 724.474955] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aba7f87a-a068-4ce9-8b1b-aaac0491e468 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.482843] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Waiting for the task: (returnval){ [ 724.482843] env[61243]: value = "task-1338755" [ 724.482843] env[61243]: _type = "Task" [ 724.482843] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.492848] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': task-1338755, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.503118] env[61243]: DEBUG nova.virt.hardware [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 724.503118] env[61243]: DEBUG nova.virt.hardware [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 724.503118] env[61243]: DEBUG nova.virt.hardware [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 724.503285] env[61243]: DEBUG nova.virt.hardware [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 724.503285] env[61243]: DEBUG nova.virt.hardware [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 724.503285] env[61243]: DEBUG nova.virt.hardware [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 724.503285] env[61243]: DEBUG nova.virt.hardware [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 724.503285] env[61243]: DEBUG nova.virt.hardware [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 724.504376] env[61243]: DEBUG nova.virt.hardware [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 724.504814] env[61243]: DEBUG nova.virt.hardware [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 724.505284] env[61243]: DEBUG nova.virt.hardware [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 724.508348] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b94b485-a9b3-43ae-a8a4-f66ec1699d92 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.517396] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66a13cf-c466-4a7d-910e-49c683de79cc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.535553] env[61243]: ERROR nova.compute.manager [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 08bf97d6-7dce-4603-9977-136dce2b29cb, please check neutron logs for more information. [ 724.535553] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Traceback (most recent call last): [ 724.535553] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 724.535553] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] yield resources [ 724.535553] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.535553] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] self.driver.spawn(context, instance, image_meta, [ 724.535553] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 724.535553] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.535553] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.535553] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] vm_ref = self.build_virtual_machine(instance, [ 724.535553] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.536248] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.536248] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.536248] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] for vif in network_info: [ 724.536248] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.536248] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] return self._sync_wrapper(fn, *args, **kwargs) [ 724.536248] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.536248] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] self.wait() [ 724.536248] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.536248] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] self[:] = self._gt.wait() [ 724.536248] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.536248] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] return self._exit_event.wait() [ 724.536248] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 724.536248] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] current.throw(*self._exc) [ 724.537397] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.537397] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] result = function(*args, **kwargs) [ 724.537397] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 724.537397] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] return func(*args, **kwargs) [ 724.537397] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.537397] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] raise e [ 724.537397] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.537397] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] nwinfo = self.network_api.allocate_for_instance( [ 724.537397] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.537397] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] created_port_ids = self._update_ports_for_instance( [ 724.537397] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.537397] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] with excutils.save_and_reraise_exception(): [ 724.537397] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.537777] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] self.force_reraise() [ 724.537777] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.537777] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] raise self.value [ 724.537777] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.537777] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] updated_port = self._update_port( [ 724.537777] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.537777] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] _ensure_no_port_binding_failure(port) [ 724.537777] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.537777] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] raise exception.PortBindingFailed(port_id=port['id']) [ 724.537777] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] nova.exception.PortBindingFailed: Binding failed for port 08bf97d6-7dce-4603-9977-136dce2b29cb, please check neutron logs for more information. [ 724.537777] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] [ 724.537777] env[61243]: INFO nova.compute.manager [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Terminating instance [ 724.539553] env[61243]: DEBUG nova.network.neutron [req-226568ec-d470-4ac5-8ebe-afdc6b5a3782 req-c856f5a4-35ae-4545-8103-bd424d89af1f service nova] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.541804] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Acquiring lock "refresh_cache-3b852a19-9f86-486b-b98d-e87ffc54d943" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.665652] env[61243]: DEBUG nova.network.neutron [req-226568ec-d470-4ac5-8ebe-afdc6b5a3782 req-c856f5a4-35ae-4545-8103-bd424d89af1f service nova] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.758436] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.892514] env[61243]: DEBUG nova.network.neutron [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.993643] env[61243]: DEBUG oslo_vmware.api [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': task-1338755, 'name': PowerOnVM_Task, 'duration_secs': 0.398071} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.993913] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 724.995025] env[61243]: INFO nova.compute.manager [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Took 4.80 seconds to spawn the instance on the hypervisor. [ 724.995025] env[61243]: DEBUG nova.compute.manager [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 724.995296] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c4473d-274f-48d8-820d-4deadf6745b5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.023222] env[61243]: DEBUG nova.network.neutron [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.168357] env[61243]: DEBUG oslo_concurrency.lockutils [req-226568ec-d470-4ac5-8ebe-afdc6b5a3782 req-c856f5a4-35ae-4545-8103-bd424d89af1f service nova] Releasing lock "refresh_cache-3b852a19-9f86-486b-b98d-e87ffc54d943" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.168882] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Acquired lock "refresh_cache-3b852a19-9f86-486b-b98d-e87ffc54d943" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.169123] env[61243]: DEBUG nova.network.neutron [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 725.518038] env[61243]: INFO nova.compute.manager [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Took 40.66 seconds to build instance. [ 725.523780] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Releasing lock "refresh_cache-c887c01b-a67e-4e3e-9e17-1deb26e9b508" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.524047] env[61243]: DEBUG nova.compute.manager [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 725.524555] env[61243]: DEBUG nova.compute.manager [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.524555] env[61243]: DEBUG nova.network.neutron [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 725.542281] env[61243]: DEBUG nova.network.neutron [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.698121] env[61243]: DEBUG nova.network.neutron [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.703175] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab36398-592f-4b4c-8b2e-495402cc0c0e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.711172] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc70e17-9c28-4f75-a6dc-19652cbe1875 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.744045] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d82dc4fe-9473-4eee-b9b9-8bb74b32afcb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.751683] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e840017a-549f-4ed5-ace6-83aee262d63b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.765457] env[61243]: DEBUG nova.compute.provider_tree [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.829679] env[61243]: DEBUG nova.network.neutron [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.021615] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d3d03afa-3fc8-4e6b-a45f-0dea714aeea4 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Lock "fd2d724a-76b6-405f-b003-54de2ad84f5a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.039s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.044731] env[61243]: DEBUG nova.network.neutron [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.068731] env[61243]: DEBUG nova.compute.manager [req-247818e1-0079-409e-97ce-f430811d4f5a req-0abc5b3f-7c85-46da-8555-781955b95bf0 service nova] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Received event network-vif-deleted-08bf97d6-7dce-4603-9977-136dce2b29cb {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 726.271018] env[61243]: DEBUG nova.scheduler.client.report [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.333794] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Releasing lock "refresh_cache-3b852a19-9f86-486b-b98d-e87ffc54d943" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.334282] env[61243]: DEBUG nova.compute.manager [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 726.334479] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 726.334776] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0df030f7-4a94-47dd-a76f-2598f160f29b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.345043] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eacacc8a-1bf5-47e2-9101-fdc5b4a95516 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.370931] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3b852a19-9f86-486b-b98d-e87ffc54d943 could not be found. [ 726.371237] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 726.371423] env[61243]: INFO nova.compute.manager [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Took 0.04 seconds to destroy the instance on the hypervisor. [ 726.371663] env[61243]: DEBUG oslo.service.loopingcall [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.372410] env[61243]: DEBUG nova.compute.manager [-] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.372605] env[61243]: DEBUG nova.network.neutron [-] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 726.388672] env[61243]: DEBUG nova.network.neutron [-] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.523185] env[61243]: DEBUG nova.compute.manager [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 726.547092] env[61243]: INFO nova.compute.manager [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] [instance: c887c01b-a67e-4e3e-9e17-1deb26e9b508] Took 1.02 seconds to deallocate network for instance. [ 726.775937] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.415s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.776402] env[61243]: DEBUG nova.compute.manager [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 726.779360] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.766s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.861579] env[61243]: DEBUG nova.compute.manager [None req-f268f9b9-1115-4c71-9494-b3f8fa24997a tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 726.862597] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709fcde9-1808-4b94-b0fe-b46330b5ddb7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.891355] env[61243]: DEBUG nova.network.neutron [-] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.939307] env[61243]: DEBUG oslo_concurrency.lockutils [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Acquiring lock "fd2d724a-76b6-405f-b003-54de2ad84f5a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.939720] env[61243]: DEBUG oslo_concurrency.lockutils [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Lock "fd2d724a-76b6-405f-b003-54de2ad84f5a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.940142] env[61243]: DEBUG oslo_concurrency.lockutils [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Acquiring lock "fd2d724a-76b6-405f-b003-54de2ad84f5a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.940788] env[61243]: DEBUG oslo_concurrency.lockutils [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Lock "fd2d724a-76b6-405f-b003-54de2ad84f5a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.941053] env[61243]: DEBUG oslo_concurrency.lockutils [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Lock "fd2d724a-76b6-405f-b003-54de2ad84f5a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.943091] env[61243]: INFO nova.compute.manager [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Terminating instance [ 726.945069] env[61243]: DEBUG oslo_concurrency.lockutils [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Acquiring lock "refresh_cache-fd2d724a-76b6-405f-b003-54de2ad84f5a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.945248] env[61243]: DEBUG oslo_concurrency.lockutils [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Acquired lock "refresh_cache-fd2d724a-76b6-405f-b003-54de2ad84f5a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.945419] env[61243]: DEBUG nova.network.neutron [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 727.051320] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.285776] env[61243]: DEBUG nova.compute.utils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 727.290797] env[61243]: DEBUG nova.compute.manager [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 727.290979] env[61243]: DEBUG nova.network.neutron [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 727.356723] env[61243]: DEBUG nova.policy [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ca954014cf04ee9b367cc4461479ff9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49a6f58599cf499e87fdc9325647b648', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 727.379430] env[61243]: INFO nova.compute.manager [None req-f268f9b9-1115-4c71-9494-b3f8fa24997a tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] instance snapshotting [ 727.380183] env[61243]: DEBUG nova.objects.instance [None req-f268f9b9-1115-4c71-9494-b3f8fa24997a tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Lazy-loading 'flavor' on Instance uuid fd2d724a-76b6-405f-b003-54de2ad84f5a {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 727.396712] env[61243]: INFO nova.compute.manager [-] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Took 1.02 seconds to deallocate network for instance. [ 727.402646] env[61243]: DEBUG nova.compute.claims [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 727.404847] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.474463] env[61243]: DEBUG nova.network.neutron [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.588401] env[61243]: INFO nova.scheduler.client.report [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Deleted allocations for instance c887c01b-a67e-4e3e-9e17-1deb26e9b508 [ 727.616800] env[61243]: DEBUG nova.network.neutron [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.707017] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dc4fcc4-1f7b-4fd7-bfc9-ea0191b920e8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.715065] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b43757-f15a-4af1-8dec-11827893fa56 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.754047] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcef0bfb-fbcb-4252-bcf1-cab2fc92e9d6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.762292] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4927786-afa3-40af-a09f-8a68b8a9d223 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.779041] env[61243]: DEBUG nova.compute.provider_tree [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.791630] env[61243]: DEBUG nova.compute.manager [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 727.890060] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63245d27-0c9c-49cd-ab66-2e4ef711c9a3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.895033] env[61243]: DEBUG nova.network.neutron [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Successfully created port: 1bf74449-72ce-4393-8d7d-821b435c9b32 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 727.912213] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4350a16-46cd-4c3e-af46-b9f13c115193 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.097914] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9948d269-b05d-4097-81d9-12636b5277d1 tempest-ServerAddressesNegativeTestJSON-2113630710 tempest-ServerAddressesNegativeTestJSON-2113630710-project-member] Lock "c887c01b-a67e-4e3e-9e17-1deb26e9b508" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.549s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.123071] env[61243]: DEBUG oslo_concurrency.lockutils [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Releasing lock "refresh_cache-fd2d724a-76b6-405f-b003-54de2ad84f5a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.123071] env[61243]: DEBUG nova.compute.manager [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 728.123071] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 728.123071] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-617b9567-8eaa-498b-a441-9ad561e63fe4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.131160] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 728.131160] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f2a717a-f5b5-429e-9867-feb436ff6708 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.142020] env[61243]: DEBUG oslo_vmware.api [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Waiting for the task: (returnval){ [ 728.142020] env[61243]: value = "task-1338756" [ 728.142020] env[61243]: _type = "Task" [ 728.142020] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.153649] env[61243]: DEBUG oslo_vmware.api [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': task-1338756, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.287026] env[61243]: DEBUG nova.scheduler.client.report [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.426503] env[61243]: DEBUG nova.compute.manager [None req-f268f9b9-1115-4c71-9494-b3f8fa24997a tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Instance disappeared during snapshot {{(pid=61243) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 728.588926] env[61243]: DEBUG nova.compute.manager [None req-f268f9b9-1115-4c71-9494-b3f8fa24997a tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Found 0 images (rotation: 2) {{(pid=61243) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 728.599872] env[61243]: DEBUG nova.compute.manager [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 728.648583] env[61243]: DEBUG oslo_vmware.api [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': task-1338756, 'name': PowerOffVM_Task, 'duration_secs': 0.181229} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.649087] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 728.649298] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 728.649554] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c51fb9dc-4df2-46f7-ac86-79050f851922 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.674427] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 728.675066] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 728.675377] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Deleting the datastore file [datastore2] fd2d724a-76b6-405f-b003-54de2ad84f5a {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 728.676192] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d3119b20-6bdc-453e-a971-89cce51f611f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.685057] env[61243]: DEBUG oslo_vmware.api [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Waiting for the task: (returnval){ [ 728.685057] env[61243]: value = "task-1338758" [ 728.685057] env[61243]: _type = "Task" [ 728.685057] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.693353] env[61243]: DEBUG oslo_vmware.api [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': task-1338758, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.790727] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.011s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.791357] env[61243]: ERROR nova.compute.manager [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 88023220-3fa7-47f6-bb81-ff5e68503d8c, please check neutron logs for more information. [ 728.791357] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Traceback (most recent call last): [ 728.791357] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 728.791357] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] self.driver.spawn(context, instance, image_meta, [ 728.791357] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 728.791357] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.791357] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.791357] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] vm_ref = self.build_virtual_machine(instance, [ 728.791357] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.791357] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.791357] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.792122] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] for vif in network_info: [ 728.792122] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.792122] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] return self._sync_wrapper(fn, *args, **kwargs) [ 728.792122] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.792122] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] self.wait() [ 728.792122] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.792122] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] self[:] = self._gt.wait() [ 728.792122] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.792122] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] return self._exit_event.wait() [ 728.792122] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 728.792122] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] current.throw(*self._exc) [ 728.792122] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.792122] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] result = function(*args, **kwargs) [ 728.792799] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 728.792799] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] return func(*args, **kwargs) [ 728.792799] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.792799] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] raise e [ 728.792799] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.792799] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] nwinfo = self.network_api.allocate_for_instance( [ 728.792799] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.792799] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] created_port_ids = self._update_ports_for_instance( [ 728.792799] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.792799] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] with excutils.save_and_reraise_exception(): [ 728.792799] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.792799] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] self.force_reraise() [ 728.792799] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.794196] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] raise self.value [ 728.794196] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.794196] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] updated_port = self._update_port( [ 728.794196] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.794196] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] _ensure_no_port_binding_failure(port) [ 728.794196] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.794196] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] raise exception.PortBindingFailed(port_id=port['id']) [ 728.794196] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] nova.exception.PortBindingFailed: Binding failed for port 88023220-3fa7-47f6-bb81-ff5e68503d8c, please check neutron logs for more information. [ 728.794196] env[61243]: ERROR nova.compute.manager [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] [ 728.794196] env[61243]: DEBUG nova.compute.utils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Binding failed for port 88023220-3fa7-47f6-bb81-ff5e68503d8c, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 728.794607] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.098s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.802283] env[61243]: DEBUG nova.compute.manager [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Build of instance cceaed50-e20f-45e4-9959-29dac2ecad30 was re-scheduled: Binding failed for port 88023220-3fa7-47f6-bb81-ff5e68503d8c, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 728.802746] env[61243]: DEBUG nova.compute.manager [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 728.802981] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Acquiring lock "refresh_cache-cceaed50-e20f-45e4-9959-29dac2ecad30" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.803147] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Acquired lock "refresh_cache-cceaed50-e20f-45e4-9959-29dac2ecad30" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.803306] env[61243]: DEBUG nova.network.neutron [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 728.804955] env[61243]: DEBUG nova.compute.manager [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 728.831404] env[61243]: DEBUG nova.virt.hardware [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 728.836018] env[61243]: DEBUG nova.virt.hardware [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 728.836018] env[61243]: DEBUG nova.virt.hardware [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 728.836018] env[61243]: DEBUG nova.virt.hardware [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 728.836018] env[61243]: DEBUG nova.virt.hardware [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 728.836018] env[61243]: DEBUG nova.virt.hardware [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 728.836243] env[61243]: DEBUG nova.virt.hardware [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 728.836243] env[61243]: DEBUG nova.virt.hardware [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 728.836243] env[61243]: DEBUG nova.virt.hardware [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 728.836243] env[61243]: DEBUG nova.virt.hardware [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 728.836243] env[61243]: DEBUG nova.virt.hardware [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 728.836388] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af220a81-9149-4fa9-a954-bc3ff9702c18 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.845995] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf66b220-27c5-44fd-9c52-4d15e40c0639 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.128399] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.193160] env[61243]: DEBUG oslo_vmware.api [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Task: {'id': task-1338758, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.416268} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.193436] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 729.197019] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 729.197019] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 729.197019] env[61243]: INFO nova.compute.manager [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Took 1.07 seconds to destroy the instance on the hypervisor. [ 729.197019] env[61243]: DEBUG oslo.service.loopingcall [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 729.197019] env[61243]: DEBUG nova.compute.manager [-] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 729.197019] env[61243]: DEBUG nova.network.neutron [-] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 729.223963] env[61243]: DEBUG nova.network.neutron [-] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.343227] env[61243]: DEBUG nova.network.neutron [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.574609] env[61243]: DEBUG nova.network.neutron [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.687082] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-326faf0e-6b9e-4551-9433-f5be64f4f0e6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.694714] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9f40321-4cea-4cfe-bfec-271437681379 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.732850] env[61243]: DEBUG nova.network.neutron [-] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.734789] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3948eb22-3ecf-4653-9011-bc2fbf3ee285 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.747447] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f0ddfd6-d2c3-48f7-8f21-92f534ef89ae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.761045] env[61243]: DEBUG nova.compute.provider_tree [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.077767] env[61243]: DEBUG nova.compute.manager [req-7371d6ac-e768-4302-ad33-3b10f2046318 req-34d0000a-1243-4d48-ad23-7b783fa969bc service nova] [instance: be8ac265-b214-4db5-9781-d171cee46538] Received event network-changed-1bf74449-72ce-4393-8d7d-821b435c9b32 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 730.077966] env[61243]: DEBUG nova.compute.manager [req-7371d6ac-e768-4302-ad33-3b10f2046318 req-34d0000a-1243-4d48-ad23-7b783fa969bc service nova] [instance: be8ac265-b214-4db5-9781-d171cee46538] Refreshing instance network info cache due to event network-changed-1bf74449-72ce-4393-8d7d-821b435c9b32. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 730.078182] env[61243]: DEBUG oslo_concurrency.lockutils [req-7371d6ac-e768-4302-ad33-3b10f2046318 req-34d0000a-1243-4d48-ad23-7b783fa969bc service nova] Acquiring lock "refresh_cache-be8ac265-b214-4db5-9781-d171cee46538" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.078339] env[61243]: DEBUG oslo_concurrency.lockutils [req-7371d6ac-e768-4302-ad33-3b10f2046318 req-34d0000a-1243-4d48-ad23-7b783fa969bc service nova] Acquired lock "refresh_cache-be8ac265-b214-4db5-9781-d171cee46538" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.078565] env[61243]: DEBUG nova.network.neutron [req-7371d6ac-e768-4302-ad33-3b10f2046318 req-34d0000a-1243-4d48-ad23-7b783fa969bc service nova] [instance: be8ac265-b214-4db5-9781-d171cee46538] Refreshing network info cache for port 1bf74449-72ce-4393-8d7d-821b435c9b32 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 730.083674] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Releasing lock "refresh_cache-cceaed50-e20f-45e4-9959-29dac2ecad30" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.083890] env[61243]: DEBUG nova.compute.manager [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 730.084098] env[61243]: DEBUG nova.compute.manager [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 730.084269] env[61243]: DEBUG nova.network.neutron [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 730.125790] env[61243]: DEBUG nova.network.neutron [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.240571] env[61243]: INFO nova.compute.manager [-] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Took 1.05 seconds to deallocate network for instance. [ 730.264895] env[61243]: DEBUG nova.scheduler.client.report [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.622476] env[61243]: DEBUG nova.network.neutron [req-7371d6ac-e768-4302-ad33-3b10f2046318 req-34d0000a-1243-4d48-ad23-7b783fa969bc service nova] [instance: be8ac265-b214-4db5-9781-d171cee46538] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.628975] env[61243]: DEBUG nova.network.neutron [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.691739] env[61243]: ERROR nova.compute.manager [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1bf74449-72ce-4393-8d7d-821b435c9b32, please check neutron logs for more information. [ 730.691739] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 730.691739] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.691739] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 730.691739] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.691739] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 730.691739] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.691739] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 730.691739] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.691739] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 730.691739] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.691739] env[61243]: ERROR nova.compute.manager raise self.value [ 730.691739] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.691739] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 730.691739] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.691739] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 730.692425] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.692425] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 730.692425] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1bf74449-72ce-4393-8d7d-821b435c9b32, please check neutron logs for more information. [ 730.692425] env[61243]: ERROR nova.compute.manager [ 730.692425] env[61243]: Traceback (most recent call last): [ 730.692425] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 730.692425] env[61243]: listener.cb(fileno) [ 730.692425] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.692425] env[61243]: result = function(*args, **kwargs) [ 730.692425] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 730.692425] env[61243]: return func(*args, **kwargs) [ 730.692425] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 730.692425] env[61243]: raise e [ 730.692425] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.692425] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 730.692425] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.692425] env[61243]: created_port_ids = self._update_ports_for_instance( [ 730.692425] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.692425] env[61243]: with excutils.save_and_reraise_exception(): [ 730.692425] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.692425] env[61243]: self.force_reraise() [ 730.692425] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.692425] env[61243]: raise self.value [ 730.692425] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.692425] env[61243]: updated_port = self._update_port( [ 730.692425] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.692425] env[61243]: _ensure_no_port_binding_failure(port) [ 730.692425] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.692425] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 730.693222] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 1bf74449-72ce-4393-8d7d-821b435c9b32, please check neutron logs for more information. [ 730.693222] env[61243]: Removing descriptor: 16 [ 730.693222] env[61243]: ERROR nova.compute.manager [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1bf74449-72ce-4393-8d7d-821b435c9b32, please check neutron logs for more information. [ 730.693222] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] Traceback (most recent call last): [ 730.693222] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 730.693222] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] yield resources [ 730.693222] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 730.693222] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] self.driver.spawn(context, instance, image_meta, [ 730.693222] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 730.693222] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.693222] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.693222] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] vm_ref = self.build_virtual_machine(instance, [ 730.693616] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.693616] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.693616] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.693616] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] for vif in network_info: [ 730.693616] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.693616] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] return self._sync_wrapper(fn, *args, **kwargs) [ 730.693616] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.693616] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] self.wait() [ 730.693616] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.693616] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] self[:] = self._gt.wait() [ 730.693616] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.693616] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] return self._exit_event.wait() [ 730.693616] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 730.694214] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] result = hub.switch() [ 730.694214] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 730.694214] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] return self.greenlet.switch() [ 730.694214] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.694214] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] result = function(*args, **kwargs) [ 730.694214] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 730.694214] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] return func(*args, **kwargs) [ 730.694214] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 730.694214] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] raise e [ 730.694214] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.694214] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] nwinfo = self.network_api.allocate_for_instance( [ 730.694214] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.694214] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] created_port_ids = self._update_ports_for_instance( [ 730.694607] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.694607] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] with excutils.save_and_reraise_exception(): [ 730.694607] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.694607] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] self.force_reraise() [ 730.694607] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.694607] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] raise self.value [ 730.694607] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.694607] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] updated_port = self._update_port( [ 730.694607] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.694607] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] _ensure_no_port_binding_failure(port) [ 730.694607] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.694607] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] raise exception.PortBindingFailed(port_id=port['id']) [ 730.694951] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] nova.exception.PortBindingFailed: Binding failed for port 1bf74449-72ce-4393-8d7d-821b435c9b32, please check neutron logs for more information. [ 730.694951] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] [ 730.694951] env[61243]: INFO nova.compute.manager [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Terminating instance [ 730.695989] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Acquiring lock "refresh_cache-be8ac265-b214-4db5-9781-d171cee46538" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.750088] env[61243]: DEBUG oslo_concurrency.lockutils [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.774693] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.980s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.775226] env[61243]: ERROR nova.compute.manager [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 304f43ea-8b5b-46a6-9a7c-a355d8f4e546, please check neutron logs for more information. [ 730.775226] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Traceback (most recent call last): [ 730.775226] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 730.775226] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] self.driver.spawn(context, instance, image_meta, [ 730.775226] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 730.775226] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.775226] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.775226] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] vm_ref = self.build_virtual_machine(instance, [ 730.775226] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.775226] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.775226] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.775543] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] for vif in network_info: [ 730.775543] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.775543] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] return self._sync_wrapper(fn, *args, **kwargs) [ 730.775543] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.775543] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] self.wait() [ 730.775543] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.775543] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] self[:] = self._gt.wait() [ 730.775543] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.775543] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] return self._exit_event.wait() [ 730.775543] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 730.775543] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] current.throw(*self._exc) [ 730.775543] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.775543] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] result = function(*args, **kwargs) [ 730.775905] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 730.775905] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] return func(*args, **kwargs) [ 730.775905] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 730.775905] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] raise e [ 730.775905] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.775905] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] nwinfo = self.network_api.allocate_for_instance( [ 730.775905] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.775905] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] created_port_ids = self._update_ports_for_instance( [ 730.775905] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.775905] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] with excutils.save_and_reraise_exception(): [ 730.775905] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.775905] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] self.force_reraise() [ 730.775905] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.776368] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] raise self.value [ 730.776368] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.776368] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] updated_port = self._update_port( [ 730.776368] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.776368] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] _ensure_no_port_binding_failure(port) [ 730.776368] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.776368] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] raise exception.PortBindingFailed(port_id=port['id']) [ 730.776368] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] nova.exception.PortBindingFailed: Binding failed for port 304f43ea-8b5b-46a6-9a7c-a355d8f4e546, please check neutron logs for more information. [ 730.776368] env[61243]: ERROR nova.compute.manager [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] [ 730.776368] env[61243]: DEBUG nova.compute.utils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Binding failed for port 304f43ea-8b5b-46a6-9a7c-a355d8f4e546, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 730.778287] env[61243]: DEBUG nova.compute.manager [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Build of instance 755cdc16-0eae-4978-917f-634a8ac7af8f was re-scheduled: Binding failed for port 304f43ea-8b5b-46a6-9a7c-a355d8f4e546, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 730.778680] env[61243]: DEBUG nova.compute.manager [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 730.778918] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Acquiring lock "refresh_cache-755cdc16-0eae-4978-917f-634a8ac7af8f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.779085] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Acquired lock "refresh_cache-755cdc16-0eae-4978-917f-634a8ac7af8f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.779445] env[61243]: DEBUG nova.network.neutron [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.780742] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.686s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.817525] env[61243]: DEBUG nova.network.neutron [req-7371d6ac-e768-4302-ad33-3b10f2046318 req-34d0000a-1243-4d48-ad23-7b783fa969bc service nova] [instance: be8ac265-b214-4db5-9781-d171cee46538] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.134116] env[61243]: INFO nova.compute.manager [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] [instance: cceaed50-e20f-45e4-9959-29dac2ecad30] Took 1.05 seconds to deallocate network for instance. [ 731.306055] env[61243]: DEBUG nova.network.neutron [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.324957] env[61243]: DEBUG oslo_concurrency.lockutils [req-7371d6ac-e768-4302-ad33-3b10f2046318 req-34d0000a-1243-4d48-ad23-7b783fa969bc service nova] Releasing lock "refresh_cache-be8ac265-b214-4db5-9781-d171cee46538" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.327541] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Acquired lock "refresh_cache-be8ac265-b214-4db5-9781-d171cee46538" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.327971] env[61243]: DEBUG nova.network.neutron [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 731.383991] env[61243]: DEBUG nova.network.neutron [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.702173] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e34e5f35-c8db-436e-ad4b-469644a42a44 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.709300] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32404c75-94ab-4d3b-9312-e0342548698b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.740928] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb0b4a1-f84e-433e-b0dc-c375a72c3a22 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.750191] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a2d994-c868-4f6a-a020-f3e990d6d12e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.766090] env[61243]: DEBUG nova.compute.provider_tree [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.854903] env[61243]: DEBUG nova.network.neutron [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.887022] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Releasing lock "refresh_cache-755cdc16-0eae-4978-917f-634a8ac7af8f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.887022] env[61243]: DEBUG nova.compute.manager [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 731.887022] env[61243]: DEBUG nova.compute.manager [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 731.887022] env[61243]: DEBUG nova.network.neutron [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 731.915533] env[61243]: DEBUG nova.network.neutron [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.944643] env[61243]: DEBUG nova.network.neutron [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.163037] env[61243]: INFO nova.scheduler.client.report [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Deleted allocations for instance cceaed50-e20f-45e4-9959-29dac2ecad30 [ 732.179893] env[61243]: DEBUG nova.compute.manager [req-918733eb-7bda-499d-be84-bbb8688a50fd req-db03bb31-3612-45ce-a18d-bf8747fc64c1 service nova] [instance: be8ac265-b214-4db5-9781-d171cee46538] Received event network-vif-deleted-1bf74449-72ce-4393-8d7d-821b435c9b32 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 732.267813] env[61243]: DEBUG nova.scheduler.client.report [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 732.418737] env[61243]: DEBUG nova.network.neutron [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.448935] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Releasing lock "refresh_cache-be8ac265-b214-4db5-9781-d171cee46538" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.449422] env[61243]: DEBUG nova.compute.manager [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 732.449599] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 732.449898] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa1da7b8-46a2-40f3-90ac-1a0a20ca0731 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.459386] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92938523-86fe-448c-9221-0eeba487d1d5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.481659] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance be8ac265-b214-4db5-9781-d171cee46538 could not be found. [ 732.481894] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 732.482196] env[61243]: INFO nova.compute.manager [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Took 0.03 seconds to destroy the instance on the hypervisor. [ 732.482387] env[61243]: DEBUG oslo.service.loopingcall [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.482609] env[61243]: DEBUG nova.compute.manager [-] [instance: be8ac265-b214-4db5-9781-d171cee46538] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 732.482704] env[61243]: DEBUG nova.network.neutron [-] [instance: be8ac265-b214-4db5-9781-d171cee46538] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 732.500153] env[61243]: DEBUG nova.network.neutron [-] [instance: be8ac265-b214-4db5-9781-d171cee46538] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.637086] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.637389] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.670363] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cb339d8d-f0e8-43e3-be3e-cda2ca74ac38 tempest-ServerPasswordTestJSON-1787021210 tempest-ServerPasswordTestJSON-1787021210-project-member] Lock "cceaed50-e20f-45e4-9959-29dac2ecad30" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 163.762s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.772846] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.991s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.773145] env[61243]: ERROR nova.compute.manager [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f8f5b1a3-8bfd-4424-9c46-7497097c2e2c, please check neutron logs for more information. [ 732.773145] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Traceback (most recent call last): [ 732.773145] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.773145] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] self.driver.spawn(context, instance, image_meta, [ 732.773145] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 732.773145] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.773145] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.773145] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] vm_ref = self.build_virtual_machine(instance, [ 732.773145] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.773145] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.773145] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.773521] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] for vif in network_info: [ 732.773521] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.773521] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] return self._sync_wrapper(fn, *args, **kwargs) [ 732.773521] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.773521] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] self.wait() [ 732.773521] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.773521] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] self[:] = self._gt.wait() [ 732.773521] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.773521] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] return self._exit_event.wait() [ 732.773521] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 732.773521] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] current.throw(*self._exc) [ 732.773521] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.773521] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] result = function(*args, **kwargs) [ 732.773912] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.773912] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] return func(*args, **kwargs) [ 732.773912] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.773912] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] raise e [ 732.773912] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.773912] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] nwinfo = self.network_api.allocate_for_instance( [ 732.773912] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.773912] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] created_port_ids = self._update_ports_for_instance( [ 732.773912] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.773912] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] with excutils.save_and_reraise_exception(): [ 732.773912] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.773912] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] self.force_reraise() [ 732.773912] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.774287] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] raise self.value [ 732.774287] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.774287] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] updated_port = self._update_port( [ 732.774287] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.774287] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] _ensure_no_port_binding_failure(port) [ 732.774287] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.774287] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] raise exception.PortBindingFailed(port_id=port['id']) [ 732.774287] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] nova.exception.PortBindingFailed: Binding failed for port f8f5b1a3-8bfd-4424-9c46-7497097c2e2c, please check neutron logs for more information. [ 732.774287] env[61243]: ERROR nova.compute.manager [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] [ 732.774287] env[61243]: DEBUG nova.compute.utils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Binding failed for port f8f5b1a3-8bfd-4424-9c46-7497097c2e2c, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 732.774897] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.194s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.778435] env[61243]: DEBUG nova.compute.manager [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Build of instance 37b43ee9-f991-4d56-8fcc-192f6534f803 was re-scheduled: Binding failed for port f8f5b1a3-8bfd-4424-9c46-7497097c2e2c, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 732.778854] env[61243]: DEBUG nova.compute.manager [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 732.779222] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Acquiring lock "refresh_cache-37b43ee9-f991-4d56-8fcc-192f6534f803" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.779478] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Acquired lock "refresh_cache-37b43ee9-f991-4d56-8fcc-192f6534f803" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.779678] env[61243]: DEBUG nova.network.neutron [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.921581] env[61243]: INFO nova.compute.manager [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] [instance: 755cdc16-0eae-4978-917f-634a8ac7af8f] Took 1.03 seconds to deallocate network for instance. [ 733.003701] env[61243]: DEBUG nova.network.neutron [-] [instance: be8ac265-b214-4db5-9781-d171cee46538] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.174093] env[61243]: DEBUG nova.compute.manager [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 733.306806] env[61243]: DEBUG nova.network.neutron [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.446110] env[61243]: DEBUG nova.network.neutron [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.505775] env[61243]: INFO nova.compute.manager [-] [instance: be8ac265-b214-4db5-9781-d171cee46538] Took 1.02 seconds to deallocate network for instance. [ 733.507601] env[61243]: DEBUG nova.compute.claims [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 733.508550] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.604064] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b22c2c1-5d6a-4f37-8ece-6d9d533b1f23 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.614639] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquiring lock "6cee3dd1-c13d-427d-889c-6cd3ae388a31" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.615059] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Lock "6cee3dd1-c13d-427d-889c-6cd3ae388a31" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.616176] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4efe9bd9-289d-43e3-b5dc-30913126a3a6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.646988] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-004214ab-b103-45f7-a52c-b3c4b1016f56 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.654734] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99e0d284-9abc-42a2-9595-d34eab115ab9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.669210] env[61243]: DEBUG nova.compute.provider_tree [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.692169] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.946399] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Releasing lock "refresh_cache-37b43ee9-f991-4d56-8fcc-192f6534f803" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.946817] env[61243]: DEBUG nova.compute.manager [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 733.947059] env[61243]: DEBUG nova.compute.manager [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.947248] env[61243]: DEBUG nova.network.neutron [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.954782] env[61243]: INFO nova.scheduler.client.report [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Deleted allocations for instance 755cdc16-0eae-4978-917f-634a8ac7af8f [ 733.964875] env[61243]: DEBUG nova.network.neutron [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.173343] env[61243]: DEBUG nova.scheduler.client.report [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.463863] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d03f91a1-089b-45c8-8378-134982cdf375 tempest-DeleteServersAdminTestJSON-210811669 tempest-DeleteServersAdminTestJSON-210811669-project-member] Lock "755cdc16-0eae-4978-917f-634a8ac7af8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.086s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.469263] env[61243]: DEBUG nova.network.neutron [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.679071] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.904s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.679702] env[61243]: ERROR nova.compute.manager [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4e9d0e19-cbce-4191-a741-c9a8b8632071, please check neutron logs for more information. [ 734.679702] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Traceback (most recent call last): [ 734.679702] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 734.679702] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] self.driver.spawn(context, instance, image_meta, [ 734.679702] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 734.679702] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.679702] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.679702] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] vm_ref = self.build_virtual_machine(instance, [ 734.679702] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.679702] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.679702] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.680080] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] for vif in network_info: [ 734.680080] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 734.680080] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] return self._sync_wrapper(fn, *args, **kwargs) [ 734.680080] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 734.680080] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] self.wait() [ 734.680080] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 734.680080] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] self[:] = self._gt.wait() [ 734.680080] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.680080] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] return self._exit_event.wait() [ 734.680080] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 734.680080] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] current.throw(*self._exc) [ 734.680080] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.680080] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] result = function(*args, **kwargs) [ 734.680556] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.680556] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] return func(*args, **kwargs) [ 734.680556] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.680556] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] raise e [ 734.680556] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.680556] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] nwinfo = self.network_api.allocate_for_instance( [ 734.680556] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.680556] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] created_port_ids = self._update_ports_for_instance( [ 734.680556] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.680556] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] with excutils.save_and_reraise_exception(): [ 734.680556] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.680556] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] self.force_reraise() [ 734.680556] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.680963] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] raise self.value [ 734.680963] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.680963] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] updated_port = self._update_port( [ 734.680963] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.680963] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] _ensure_no_port_binding_failure(port) [ 734.680963] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.680963] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] raise exception.PortBindingFailed(port_id=port['id']) [ 734.680963] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] nova.exception.PortBindingFailed: Binding failed for port 4e9d0e19-cbce-4191-a741-c9a8b8632071, please check neutron logs for more information. [ 734.680963] env[61243]: ERROR nova.compute.manager [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] [ 734.680963] env[61243]: DEBUG nova.compute.utils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Binding failed for port 4e9d0e19-cbce-4191-a741-c9a8b8632071, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 734.682088] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.596s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.684860] env[61243]: DEBUG nova.compute.manager [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Build of instance 8e85f285-1190-4c15-bc1a-8364434dc4ba was re-scheduled: Binding failed for port 4e9d0e19-cbce-4191-a741-c9a8b8632071, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 734.685299] env[61243]: DEBUG nova.compute.manager [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 734.685529] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Acquiring lock "refresh_cache-8e85f285-1190-4c15-bc1a-8364434dc4ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.685678] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Acquired lock "refresh_cache-8e85f285-1190-4c15-bc1a-8364434dc4ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.685836] env[61243]: DEBUG nova.network.neutron [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 734.966519] env[61243]: DEBUG nova.compute.manager [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 734.973091] env[61243]: INFO nova.compute.manager [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] [instance: 37b43ee9-f991-4d56-8fcc-192f6534f803] Took 1.03 seconds to deallocate network for instance. [ 735.207393] env[61243]: DEBUG nova.network.neutron [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.305579] env[61243]: DEBUG nova.network.neutron [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.483278] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.484791] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a62b9c-d6e9-46d3-8628-c01bdc96240b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.492790] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4bb8a3-4fe8-4fbc-8187-c3e94832c206 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.523180] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d851acd-72bf-4585-8b83-25f24245264d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.531500] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff86a9d-9bad-4ec6-bd3d-0a2d59ff22f9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.545277] env[61243]: DEBUG nova.compute.provider_tree [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.808697] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Releasing lock "refresh_cache-8e85f285-1190-4c15-bc1a-8364434dc4ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.808914] env[61243]: DEBUG nova.compute.manager [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 735.809154] env[61243]: DEBUG nova.compute.manager [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.809288] env[61243]: DEBUG nova.network.neutron [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.830913] env[61243]: DEBUG nova.network.neutron [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.005604] env[61243]: INFO nova.scheduler.client.report [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Deleted allocations for instance 37b43ee9-f991-4d56-8fcc-192f6534f803 [ 736.048380] env[61243]: DEBUG nova.scheduler.client.report [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 736.334337] env[61243]: DEBUG nova.network.neutron [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.514029] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12faecd7-15cf-4a35-9d4c-f9d6522eefb4 tempest-ServerGroupTestJSON-867371837 tempest-ServerGroupTestJSON-867371837-project-member] Lock "37b43ee9-f991-4d56-8fcc-192f6534f803" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.424s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.552892] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.871s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.553578] env[61243]: ERROR nova.compute.manager [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 779de90f-7a8d-43b5-a1ba-b24777560cfd, please check neutron logs for more information. [ 736.553578] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Traceback (most recent call last): [ 736.553578] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 736.553578] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] self.driver.spawn(context, instance, image_meta, [ 736.553578] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 736.553578] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] self._vmops.spawn(context, instance, image_meta, injected_files, [ 736.553578] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 736.553578] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] vm_ref = self.build_virtual_machine(instance, [ 736.553578] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 736.553578] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] vif_infos = vmwarevif.get_vif_info(self._session, [ 736.553578] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 736.553943] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] for vif in network_info: [ 736.553943] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 736.553943] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] return self._sync_wrapper(fn, *args, **kwargs) [ 736.553943] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 736.553943] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] self.wait() [ 736.553943] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 736.553943] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] self[:] = self._gt.wait() [ 736.553943] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 736.553943] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] return self._exit_event.wait() [ 736.553943] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 736.553943] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] current.throw(*self._exc) [ 736.553943] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.553943] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] result = function(*args, **kwargs) [ 736.554354] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 736.554354] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] return func(*args, **kwargs) [ 736.554354] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 736.554354] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] raise e [ 736.554354] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.554354] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] nwinfo = self.network_api.allocate_for_instance( [ 736.554354] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.554354] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] created_port_ids = self._update_ports_for_instance( [ 736.554354] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.554354] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] with excutils.save_and_reraise_exception(): [ 736.554354] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.554354] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] self.force_reraise() [ 736.554354] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.554731] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] raise self.value [ 736.554731] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.554731] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] updated_port = self._update_port( [ 736.554731] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.554731] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] _ensure_no_port_binding_failure(port) [ 736.554731] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.554731] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] raise exception.PortBindingFailed(port_id=port['id']) [ 736.554731] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] nova.exception.PortBindingFailed: Binding failed for port 779de90f-7a8d-43b5-a1ba-b24777560cfd, please check neutron logs for more information. [ 736.554731] env[61243]: ERROR nova.compute.manager [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] [ 736.554731] env[61243]: DEBUG nova.compute.utils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Binding failed for port 779de90f-7a8d-43b5-a1ba-b24777560cfd, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 736.555609] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.797s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.557146] env[61243]: INFO nova.compute.claims [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 736.560108] env[61243]: DEBUG nova.compute.manager [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Build of instance f6566092-9921-4a94-a0b5-a2d3803a9d18 was re-scheduled: Binding failed for port 779de90f-7a8d-43b5-a1ba-b24777560cfd, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 736.560220] env[61243]: DEBUG nova.compute.manager [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 736.560428] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Acquiring lock "refresh_cache-f6566092-9921-4a94-a0b5-a2d3803a9d18" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.560580] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Acquired lock "refresh_cache-f6566092-9921-4a94-a0b5-a2d3803a9d18" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.560738] env[61243]: DEBUG nova.network.neutron [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 736.840120] env[61243]: INFO nova.compute.manager [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: 8e85f285-1190-4c15-bc1a-8364434dc4ba] Took 1.03 seconds to deallocate network for instance. [ 737.016637] env[61243]: DEBUG nova.compute.manager [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 737.083223] env[61243]: DEBUG nova.network.neutron [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.167829] env[61243]: DEBUG nova.network.neutron [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.539185] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.672988] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Releasing lock "refresh_cache-f6566092-9921-4a94-a0b5-a2d3803a9d18" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.673276] env[61243]: DEBUG nova.compute.manager [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 737.673463] env[61243]: DEBUG nova.compute.manager [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 737.673634] env[61243]: DEBUG nova.network.neutron [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 737.687835] env[61243]: DEBUG nova.network.neutron [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.866980] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca2e8f9-eddb-4f19-b4f0-21135596c4c9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.870688] env[61243]: INFO nova.scheduler.client.report [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Deleted allocations for instance 8e85f285-1190-4c15-bc1a-8364434dc4ba [ 737.881770] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62b329b-76ff-4fb8-a7c1-4ee41e20d17a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.915258] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-004f4752-e43a-4a71-8097-5fa5ccbf1a4b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.923526] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a05c34dd-136a-4d37-8150-ef59f82e9cc2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.939410] env[61243]: DEBUG nova.compute.provider_tree [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.193125] env[61243]: DEBUG nova.network.neutron [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.386048] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c2fbe833-8e04-456b-9037-0c64ff294b0a tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Lock "8e85f285-1190-4c15-bc1a-8364434dc4ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.320s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.442306] env[61243]: DEBUG nova.scheduler.client.report [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 738.698392] env[61243]: INFO nova.compute.manager [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] [instance: f6566092-9921-4a94-a0b5-a2d3803a9d18] Took 1.02 seconds to deallocate network for instance. [ 738.888633] env[61243]: DEBUG nova.compute.manager [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 738.946832] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.391s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.947254] env[61243]: DEBUG nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 738.949996] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.899s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.951497] env[61243]: INFO nova.compute.claims [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 739.414984] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.455616] env[61243]: DEBUG nova.compute.utils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 739.458588] env[61243]: DEBUG nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 739.458751] env[61243]: DEBUG nova.network.neutron [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 739.502490] env[61243]: DEBUG nova.policy [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed14b085c474440982533159319b36a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b79ac625d52c4ac8b244ce45bd90b7ed', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 739.727497] env[61243]: INFO nova.scheduler.client.report [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Deleted allocations for instance f6566092-9921-4a94-a0b5-a2d3803a9d18 [ 739.767306] env[61243]: DEBUG nova.network.neutron [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Successfully created port: 54ba2a00-34fa-4c51-941a-8b4175edb70c {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 739.964383] env[61243]: DEBUG nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 740.231026] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54382a7-f128-4059-9eac-2685f215a5c3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.238692] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b3de54c9-85f6-4f3f-87c3-df68392dc866 tempest-ServerRescueNegativeTestJSON-1386564824 tempest-ServerRescueNegativeTestJSON-1386564824-project-member] Lock "f6566092-9921-4a94-a0b5-a2d3803a9d18" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 162.146s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.241186] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21440848-68f3-43c2-bbcf-54458037b1be {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.279669] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dce89ebe-a6d8-4901-adda-db3174b51adc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.288303] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7450949-aa26-46ee-9c28-d272bbf7afc9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.302582] env[61243]: DEBUG nova.compute.provider_tree [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.472324] env[61243]: INFO nova.virt.block_device [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Booting with volume 71112285-22c2-4e86-a5fb-c27b2d71ed2e at /dev/sda [ 740.518811] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7890e9e2-7e07-400f-89c8-76b9b8656c67 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.529078] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6480f81-da61-4398-b5aa-3d5e0576ea8c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.551683] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-88c1db16-3f35-48ad-bcb5-8cf47565b47b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.562743] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99102bce-44ea-415d-8cd4-f0247a5980f2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.572940] env[61243]: DEBUG nova.compute.manager [req-0148ccd3-e390-44c2-b5ff-d99bd8565829 req-7f29b1fe-9e40-4cb0-8adb-1e76799c4b88 service nova] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Received event network-changed-54ba2a00-34fa-4c51-941a-8b4175edb70c {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.573140] env[61243]: DEBUG nova.compute.manager [req-0148ccd3-e390-44c2-b5ff-d99bd8565829 req-7f29b1fe-9e40-4cb0-8adb-1e76799c4b88 service nova] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Refreshing instance network info cache due to event network-changed-54ba2a00-34fa-4c51-941a-8b4175edb70c. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 740.573378] env[61243]: DEBUG oslo_concurrency.lockutils [req-0148ccd3-e390-44c2-b5ff-d99bd8565829 req-7f29b1fe-9e40-4cb0-8adb-1e76799c4b88 service nova] Acquiring lock "refresh_cache-e02f41be-125a-4228-8136-9e6639a268a9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.573533] env[61243]: DEBUG oslo_concurrency.lockutils [req-0148ccd3-e390-44c2-b5ff-d99bd8565829 req-7f29b1fe-9e40-4cb0-8adb-1e76799c4b88 service nova] Acquired lock "refresh_cache-e02f41be-125a-4228-8136-9e6639a268a9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.573692] env[61243]: DEBUG nova.network.neutron [req-0148ccd3-e390-44c2-b5ff-d99bd8565829 req-7f29b1fe-9e40-4cb0-8adb-1e76799c4b88 service nova] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Refreshing network info cache for port 54ba2a00-34fa-4c51-941a-8b4175edb70c {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 740.589635] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b8c248-7b3d-49c3-b91b-0dc19ab93faf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.596457] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67738d44-da3c-4418-80df-e3009d005576 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.616103] env[61243]: DEBUG nova.virt.block_device [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Updating existing volume attachment record: 3690707f-f3fb-4d0f-9b09-72f8edd88afd {{(pid=61243) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 740.741892] env[61243]: DEBUG nova.compute.manager [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 740.809244] env[61243]: DEBUG nova.scheduler.client.report [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 740.987580] env[61243]: ERROR nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 54ba2a00-34fa-4c51-941a-8b4175edb70c, please check neutron logs for more information. [ 740.987580] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 740.987580] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.987580] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 740.987580] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.987580] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 740.987580] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.987580] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 740.987580] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.987580] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 740.987580] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.987580] env[61243]: ERROR nova.compute.manager raise self.value [ 740.987580] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.987580] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 740.987580] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.987580] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 740.988332] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.988332] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 740.988332] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 54ba2a00-34fa-4c51-941a-8b4175edb70c, please check neutron logs for more information. [ 740.988332] env[61243]: ERROR nova.compute.manager [ 740.988332] env[61243]: Traceback (most recent call last): [ 740.988332] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 740.988332] env[61243]: listener.cb(fileno) [ 740.988332] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.988332] env[61243]: result = function(*args, **kwargs) [ 740.988332] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 740.988332] env[61243]: return func(*args, **kwargs) [ 740.988332] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.988332] env[61243]: raise e [ 740.988332] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.988332] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 740.988332] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.988332] env[61243]: created_port_ids = self._update_ports_for_instance( [ 740.988332] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.988332] env[61243]: with excutils.save_and_reraise_exception(): [ 740.988332] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.988332] env[61243]: self.force_reraise() [ 740.988332] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.988332] env[61243]: raise self.value [ 740.988332] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.988332] env[61243]: updated_port = self._update_port( [ 740.988332] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.988332] env[61243]: _ensure_no_port_binding_failure(port) [ 740.988332] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.988332] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 740.989167] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 54ba2a00-34fa-4c51-941a-8b4175edb70c, please check neutron logs for more information. [ 740.989167] env[61243]: Removing descriptor: 15 [ 741.101891] env[61243]: DEBUG nova.network.neutron [req-0148ccd3-e390-44c2-b5ff-d99bd8565829 req-7f29b1fe-9e40-4cb0-8adb-1e76799c4b88 service nova] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.226209] env[61243]: DEBUG nova.network.neutron [req-0148ccd3-e390-44c2-b5ff-d99bd8565829 req-7f29b1fe-9e40-4cb0-8adb-1e76799c4b88 service nova] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.269113] env[61243]: DEBUG oslo_concurrency.lockutils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.314371] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.314916] env[61243]: DEBUG nova.compute.manager [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 741.317437] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.914s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.727847] env[61243]: DEBUG oslo_concurrency.lockutils [req-0148ccd3-e390-44c2-b5ff-d99bd8565829 req-7f29b1fe-9e40-4cb0-8adb-1e76799c4b88 service nova] Releasing lock "refresh_cache-e02f41be-125a-4228-8136-9e6639a268a9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.822667] env[61243]: DEBUG nova.compute.utils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 741.826823] env[61243]: DEBUG nova.compute.manager [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 741.828089] env[61243]: DEBUG nova.network.neutron [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 741.905090] env[61243]: DEBUG nova.policy [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4f62a66ae024c0f88b94b434123207c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22e2848953de459a9fcc2b78391d708c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 742.138066] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-284167da-d09f-4c25-977a-490b86b146c2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.145816] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-586d9423-9058-4792-8bb2-02cf73919809 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.175790] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18b5c095-fe37-4492-b850-b966dd03fcfc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.183867] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a08053-cffa-4d92-a8ac-6ed4b71c4e35 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.196915] env[61243]: DEBUG nova.compute.provider_tree [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.327256] env[61243]: DEBUG nova.network.neutron [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Successfully created port: 1a7dac77-793c-46dc-a760-ec1986c26daf {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.330956] env[61243]: DEBUG nova.compute.manager [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 742.655742] env[61243]: DEBUG nova.compute.manager [req-6b911984-ebc4-423a-8c1f-76063be96ce2 req-8393c124-b464-4284-bf77-4f9962db1c88 service nova] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Received event network-vif-deleted-54ba2a00-34fa-4c51-941a-8b4175edb70c {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.700708] env[61243]: DEBUG nova.scheduler.client.report [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 742.720412] env[61243]: DEBUG nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 742.720858] env[61243]: DEBUG nova.virt.hardware [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 742.720988] env[61243]: DEBUG nova.virt.hardware [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 742.721167] env[61243]: DEBUG nova.virt.hardware [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 742.721371] env[61243]: DEBUG nova.virt.hardware [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 742.721519] env[61243]: DEBUG nova.virt.hardware [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 742.721666] env[61243]: DEBUG nova.virt.hardware [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 742.721873] env[61243]: DEBUG nova.virt.hardware [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 742.722045] env[61243]: DEBUG nova.virt.hardware [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 742.722216] env[61243]: DEBUG nova.virt.hardware [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 742.722403] env[61243]: DEBUG nova.virt.hardware [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 742.722554] env[61243]: DEBUG nova.virt.hardware [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 742.728310] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1098cfc-9be9-40f5-89f2-70a9a2b6deda {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.737240] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98af00e7-b56f-40b6-9825-561270b084a2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.752739] env[61243]: ERROR nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 54ba2a00-34fa-4c51-941a-8b4175edb70c, please check neutron logs for more information. [ 742.752739] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] Traceback (most recent call last): [ 742.752739] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 742.752739] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] yield resources [ 742.752739] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 742.752739] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] self.driver.spawn(context, instance, image_meta, [ 742.752739] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 742.752739] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.752739] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.752739] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] vm_ref = self.build_virtual_machine(instance, [ 742.752739] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.753419] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.753419] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.753419] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] for vif in network_info: [ 742.753419] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 742.753419] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] return self._sync_wrapper(fn, *args, **kwargs) [ 742.753419] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 742.753419] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] self.wait() [ 742.753419] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 742.753419] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] self[:] = self._gt.wait() [ 742.753419] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.753419] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] return self._exit_event.wait() [ 742.753419] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 742.753419] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] current.throw(*self._exc) [ 742.754365] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.754365] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] result = function(*args, **kwargs) [ 742.754365] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 742.754365] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] return func(*args, **kwargs) [ 742.754365] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 742.754365] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] raise e [ 742.754365] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 742.754365] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] nwinfo = self.network_api.allocate_for_instance( [ 742.754365] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.754365] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] created_port_ids = self._update_ports_for_instance( [ 742.754365] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.754365] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] with excutils.save_and_reraise_exception(): [ 742.754365] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.755369] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] self.force_reraise() [ 742.755369] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.755369] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] raise self.value [ 742.755369] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.755369] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] updated_port = self._update_port( [ 742.755369] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.755369] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] _ensure_no_port_binding_failure(port) [ 742.755369] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.755369] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] raise exception.PortBindingFailed(port_id=port['id']) [ 742.755369] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] nova.exception.PortBindingFailed: Binding failed for port 54ba2a00-34fa-4c51-941a-8b4175edb70c, please check neutron logs for more information. [ 742.755369] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] [ 742.755369] env[61243]: INFO nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Terminating instance [ 742.756235] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Acquiring lock "refresh_cache-e02f41be-125a-4228-8136-9e6639a268a9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.756235] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Acquired lock "refresh_cache-e02f41be-125a-4228-8136-9e6639a268a9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.756235] env[61243]: DEBUG nova.network.neutron [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 743.208285] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.891s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.209048] env[61243]: ERROR nova.compute.manager [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 08bf97d6-7dce-4603-9977-136dce2b29cb, please check neutron logs for more information. [ 743.209048] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Traceback (most recent call last): [ 743.209048] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 743.209048] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] self.driver.spawn(context, instance, image_meta, [ 743.209048] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 743.209048] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] self._vmops.spawn(context, instance, image_meta, injected_files, [ 743.209048] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 743.209048] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] vm_ref = self.build_virtual_machine(instance, [ 743.209048] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 743.209048] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] vif_infos = vmwarevif.get_vif_info(self._session, [ 743.209048] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 743.209388] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] for vif in network_info: [ 743.209388] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 743.209388] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] return self._sync_wrapper(fn, *args, **kwargs) [ 743.209388] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 743.209388] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] self.wait() [ 743.209388] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 743.209388] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] self[:] = self._gt.wait() [ 743.209388] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 743.209388] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] return self._exit_event.wait() [ 743.209388] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 743.209388] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] current.throw(*self._exc) [ 743.209388] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.209388] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] result = function(*args, **kwargs) [ 743.209877] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 743.209877] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] return func(*args, **kwargs) [ 743.209877] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.209877] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] raise e [ 743.209877] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.209877] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] nwinfo = self.network_api.allocate_for_instance( [ 743.209877] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.209877] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] created_port_ids = self._update_ports_for_instance( [ 743.209877] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.209877] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] with excutils.save_and_reraise_exception(): [ 743.209877] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.209877] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] self.force_reraise() [ 743.209877] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.210276] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] raise self.value [ 743.210276] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.210276] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] updated_port = self._update_port( [ 743.210276] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.210276] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] _ensure_no_port_binding_failure(port) [ 743.210276] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.210276] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] raise exception.PortBindingFailed(port_id=port['id']) [ 743.210276] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] nova.exception.PortBindingFailed: Binding failed for port 08bf97d6-7dce-4603-9977-136dce2b29cb, please check neutron logs for more information. [ 743.210276] env[61243]: ERROR nova.compute.manager [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] [ 743.210276] env[61243]: DEBUG nova.compute.utils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Binding failed for port 08bf97d6-7dce-4603-9977-136dce2b29cb, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 743.211720] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.083s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.213223] env[61243]: INFO nova.compute.claims [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 743.215813] env[61243]: DEBUG nova.compute.manager [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Build of instance 3b852a19-9f86-486b-b98d-e87ffc54d943 was re-scheduled: Binding failed for port 08bf97d6-7dce-4603-9977-136dce2b29cb, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 743.216251] env[61243]: DEBUG nova.compute.manager [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 743.216478] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Acquiring lock "refresh_cache-3b852a19-9f86-486b-b98d-e87ffc54d943" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.216624] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Acquired lock "refresh_cache-3b852a19-9f86-486b-b98d-e87ffc54d943" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.216783] env[61243]: DEBUG nova.network.neutron [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 743.229523] env[61243]: ERROR nova.compute.manager [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1a7dac77-793c-46dc-a760-ec1986c26daf, please check neutron logs for more information. [ 743.229523] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 743.229523] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.229523] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 743.229523] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.229523] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 743.229523] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.229523] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 743.229523] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.229523] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 743.229523] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.229523] env[61243]: ERROR nova.compute.manager raise self.value [ 743.229523] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.229523] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 743.229523] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.229523] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 743.230113] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.230113] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 743.230113] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1a7dac77-793c-46dc-a760-ec1986c26daf, please check neutron logs for more information. [ 743.230113] env[61243]: ERROR nova.compute.manager [ 743.230113] env[61243]: Traceback (most recent call last): [ 743.230113] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 743.230113] env[61243]: listener.cb(fileno) [ 743.230113] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.230113] env[61243]: result = function(*args, **kwargs) [ 743.230113] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 743.230113] env[61243]: return func(*args, **kwargs) [ 743.230113] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.230113] env[61243]: raise e [ 743.230113] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.230113] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 743.230113] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.230113] env[61243]: created_port_ids = self._update_ports_for_instance( [ 743.230113] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.230113] env[61243]: with excutils.save_and_reraise_exception(): [ 743.230113] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.230113] env[61243]: self.force_reraise() [ 743.230113] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.230113] env[61243]: raise self.value [ 743.230113] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.230113] env[61243]: updated_port = self._update_port( [ 743.230113] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.230113] env[61243]: _ensure_no_port_binding_failure(port) [ 743.230113] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.230113] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 743.231034] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 1a7dac77-793c-46dc-a760-ec1986c26daf, please check neutron logs for more information. [ 743.231034] env[61243]: Removing descriptor: 15 [ 743.277646] env[61243]: DEBUG nova.network.neutron [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.338725] env[61243]: DEBUG nova.compute.manager [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 743.350208] env[61243]: DEBUG nova.network.neutron [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.366487] env[61243]: DEBUG nova.virt.hardware [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:04:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='484da1ca-225e-47c9-9755-c2f63fc26fc4',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-261334752',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 743.366729] env[61243]: DEBUG nova.virt.hardware [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 743.366907] env[61243]: DEBUG nova.virt.hardware [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 743.367133] env[61243]: DEBUG nova.virt.hardware [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 743.367287] env[61243]: DEBUG nova.virt.hardware [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 743.367436] env[61243]: DEBUG nova.virt.hardware [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 743.367643] env[61243]: DEBUG nova.virt.hardware [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 743.367800] env[61243]: DEBUG nova.virt.hardware [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 743.367961] env[61243]: DEBUG nova.virt.hardware [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 743.368151] env[61243]: DEBUG nova.virt.hardware [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 743.368327] env[61243]: DEBUG nova.virt.hardware [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 743.369369] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3e7174-c9bb-4595-b4d9-e18b6e1f24cb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.377728] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5de49e3-276a-4bf6-a608-c89b7b4552bc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.391568] env[61243]: ERROR nova.compute.manager [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1a7dac77-793c-46dc-a760-ec1986c26daf, please check neutron logs for more information. [ 743.391568] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Traceback (most recent call last): [ 743.391568] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 743.391568] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] yield resources [ 743.391568] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 743.391568] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] self.driver.spawn(context, instance, image_meta, [ 743.391568] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 743.391568] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] self._vmops.spawn(context, instance, image_meta, injected_files, [ 743.391568] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 743.391568] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] vm_ref = self.build_virtual_machine(instance, [ 743.391568] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 743.391900] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] vif_infos = vmwarevif.get_vif_info(self._session, [ 743.391900] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 743.391900] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] for vif in network_info: [ 743.391900] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 743.391900] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] return self._sync_wrapper(fn, *args, **kwargs) [ 743.391900] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 743.391900] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] self.wait() [ 743.391900] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 743.391900] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] self[:] = self._gt.wait() [ 743.391900] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 743.391900] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] return self._exit_event.wait() [ 743.391900] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 743.391900] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] current.throw(*self._exc) [ 743.392353] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.392353] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] result = function(*args, **kwargs) [ 743.392353] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 743.392353] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] return func(*args, **kwargs) [ 743.392353] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.392353] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] raise e [ 743.392353] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.392353] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] nwinfo = self.network_api.allocate_for_instance( [ 743.392353] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.392353] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] created_port_ids = self._update_ports_for_instance( [ 743.392353] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.392353] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] with excutils.save_and_reraise_exception(): [ 743.392353] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.392805] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] self.force_reraise() [ 743.392805] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.392805] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] raise self.value [ 743.392805] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.392805] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] updated_port = self._update_port( [ 743.392805] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.392805] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] _ensure_no_port_binding_failure(port) [ 743.392805] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.392805] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] raise exception.PortBindingFailed(port_id=port['id']) [ 743.392805] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] nova.exception.PortBindingFailed: Binding failed for port 1a7dac77-793c-46dc-a760-ec1986c26daf, please check neutron logs for more information. [ 743.392805] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] [ 743.392805] env[61243]: INFO nova.compute.manager [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Terminating instance [ 743.393824] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Acquiring lock "refresh_cache-b1bad82d-d7bf-4e58-bbf2-f3d68f214115" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.393982] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Acquired lock "refresh_cache-b1bad82d-d7bf-4e58-bbf2-f3d68f214115" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.394168] env[61243]: DEBUG nova.network.neutron [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 743.740726] env[61243]: DEBUG nova.network.neutron [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.814051] env[61243]: DEBUG nova.network.neutron [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.852791] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Releasing lock "refresh_cache-e02f41be-125a-4228-8136-9e6639a268a9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.853849] env[61243]: DEBUG nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 743.853849] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b8563665-187a-41fd-90a3-46a4ccc1ee83 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.862683] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4783855d-98cd-4449-9d32-cf9e04e53ae5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.883063] env[61243]: WARNING nova.virt.vmwareapi.driver [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance e02f41be-125a-4228-8136-9e6639a268a9 could not be found. [ 743.883311] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 743.883544] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b9d2744d-e8b8-4188-b364-70bd97e2f6fc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.890869] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abdc8d3a-a4ec-43aa-904b-a5167580fd9c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.914573] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e02f41be-125a-4228-8136-9e6639a268a9 could not be found. [ 743.914796] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 743.914976] env[61243]: INFO nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Took 0.06 seconds to destroy the instance on the hypervisor. [ 743.915234] env[61243]: DEBUG oslo.service.loopingcall [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 743.915450] env[61243]: DEBUG nova.compute.manager [-] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 743.915551] env[61243]: DEBUG nova.network.neutron [-] [instance: e02f41be-125a-4228-8136-9e6639a268a9] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 743.917531] env[61243]: DEBUG nova.network.neutron [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.936066] env[61243]: DEBUG nova.network.neutron [-] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.982067] env[61243]: DEBUG nova.network.neutron [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.317029] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Releasing lock "refresh_cache-3b852a19-9f86-486b-b98d-e87ffc54d943" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.317029] env[61243]: DEBUG nova.compute.manager [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 744.317273] env[61243]: DEBUG nova.compute.manager [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 744.317378] env[61243]: DEBUG nova.network.neutron [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 744.335345] env[61243]: DEBUG nova.network.neutron [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.438346] env[61243]: DEBUG nova.network.neutron [-] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.471398] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d4984f-c6d8-4f34-ad6c-6f6ceb60821a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.479136] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2beefbf4-6a45-43e8-9ec4-9aac32af6324 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.507902] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Releasing lock "refresh_cache-b1bad82d-d7bf-4e58-bbf2-f3d68f214115" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.508367] env[61243]: DEBUG nova.compute.manager [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 744.508547] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 744.510024] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-359e7090-ff32-4db3-a9bf-6ddcf4ddc8fc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.511568] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815ce02e-0861-431a-982f-bb22ba97d1c1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.520018] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d20abbb-144b-4575-aa50-eeb23ff8c613 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.526021] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4339ad25-3954-452b-8856-e52b0efa4193 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.549684] env[61243]: DEBUG nova.compute.provider_tree [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.554291] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b1bad82d-d7bf-4e58-bbf2-f3d68f214115 could not be found. [ 744.554519] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 744.554726] env[61243]: INFO nova.compute.manager [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Took 0.05 seconds to destroy the instance on the hypervisor. [ 744.554966] env[61243]: DEBUG oslo.service.loopingcall [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 744.555368] env[61243]: DEBUG nova.compute.manager [-] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 744.555467] env[61243]: DEBUG nova.network.neutron [-] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 744.570264] env[61243]: DEBUG nova.network.neutron [-] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.679851] env[61243]: DEBUG nova.compute.manager [req-75a446fc-05b3-40c9-9560-1a61e7d7066e req-fd7fc29f-c866-4228-8660-507ad259e398 service nova] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Received event network-changed-1a7dac77-793c-46dc-a760-ec1986c26daf {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 744.680054] env[61243]: DEBUG nova.compute.manager [req-75a446fc-05b3-40c9-9560-1a61e7d7066e req-fd7fc29f-c866-4228-8660-507ad259e398 service nova] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Refreshing instance network info cache due to event network-changed-1a7dac77-793c-46dc-a760-ec1986c26daf. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 744.680268] env[61243]: DEBUG oslo_concurrency.lockutils [req-75a446fc-05b3-40c9-9560-1a61e7d7066e req-fd7fc29f-c866-4228-8660-507ad259e398 service nova] Acquiring lock "refresh_cache-b1bad82d-d7bf-4e58-bbf2-f3d68f214115" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.680410] env[61243]: DEBUG oslo_concurrency.lockutils [req-75a446fc-05b3-40c9-9560-1a61e7d7066e req-fd7fc29f-c866-4228-8660-507ad259e398 service nova] Acquired lock "refresh_cache-b1bad82d-d7bf-4e58-bbf2-f3d68f214115" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.680566] env[61243]: DEBUG nova.network.neutron [req-75a446fc-05b3-40c9-9560-1a61e7d7066e req-fd7fc29f-c866-4228-8660-507ad259e398 service nova] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Refreshing network info cache for port 1a7dac77-793c-46dc-a760-ec1986c26daf {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 744.838907] env[61243]: DEBUG nova.network.neutron [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.943426] env[61243]: INFO nova.compute.manager [-] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Took 1.03 seconds to deallocate network for instance. [ 745.057059] env[61243]: DEBUG nova.scheduler.client.report [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.072013] env[61243]: DEBUG nova.network.neutron [-] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.198497] env[61243]: DEBUG nova.network.neutron [req-75a446fc-05b3-40c9-9560-1a61e7d7066e req-fd7fc29f-c866-4228-8660-507ad259e398 service nova] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.242039] env[61243]: DEBUG nova.network.neutron [req-75a446fc-05b3-40c9-9560-1a61e7d7066e req-fd7fc29f-c866-4228-8660-507ad259e398 service nova] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.341207] env[61243]: INFO nova.compute.manager [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] [instance: 3b852a19-9f86-486b-b98d-e87ffc54d943] Took 1.02 seconds to deallocate network for instance. [ 745.495411] env[61243]: INFO nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Took 0.55 seconds to detach 1 volumes for instance. [ 745.497892] env[61243]: DEBUG nova.compute.claims [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 745.498147] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.562991] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.351s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.563789] env[61243]: DEBUG nova.compute.manager [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 745.566143] env[61243]: DEBUG oslo_concurrency.lockutils [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.816s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.566342] env[61243]: DEBUG nova.objects.instance [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Lazy-loading 'resources' on Instance uuid fd2d724a-76b6-405f-b003-54de2ad84f5a {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 745.575159] env[61243]: INFO nova.compute.manager [-] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Took 1.02 seconds to deallocate network for instance. [ 745.578055] env[61243]: DEBUG nova.compute.claims [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 745.578270] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.744576] env[61243]: DEBUG oslo_concurrency.lockutils [req-75a446fc-05b3-40c9-9560-1a61e7d7066e req-fd7fc29f-c866-4228-8660-507ad259e398 service nova] Releasing lock "refresh_cache-b1bad82d-d7bf-4e58-bbf2-f3d68f214115" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.744896] env[61243]: DEBUG nova.compute.manager [req-75a446fc-05b3-40c9-9560-1a61e7d7066e req-fd7fc29f-c866-4228-8660-507ad259e398 service nova] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Received event network-vif-deleted-1a7dac77-793c-46dc-a760-ec1986c26daf {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.069966] env[61243]: DEBUG nova.compute.utils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 746.074426] env[61243]: DEBUG nova.compute.manager [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 746.074733] env[61243]: DEBUG nova.network.neutron [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 746.115695] env[61243]: DEBUG nova.policy [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7cff8e952ce1400681b834157eaf9dd6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8fbcee03fda4bff9730bf5c2a97570c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 746.305790] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa13bb89-ea33-41c4-891c-85a5bf60a8a1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.313561] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be158107-53e5-4e75-afdb-e157e6a05e0e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.343586] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c372c9-c497-4a8e-9950-63790f7295e0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.353923] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3b0ee1-fc2b-4b53-ae70-5f7cbdfd2de9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.367288] env[61243]: DEBUG nova.compute.provider_tree [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.369039] env[61243]: DEBUG nova.network.neutron [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Successfully created port: 8f597029-2d2c-4f15-8b33-593740cd82d6 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 746.371454] env[61243]: INFO nova.scheduler.client.report [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Deleted allocations for instance 3b852a19-9f86-486b-b98d-e87ffc54d943 [ 746.575733] env[61243]: DEBUG nova.compute.manager [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 746.872275] env[61243]: DEBUG nova.scheduler.client.report [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.880033] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2cc45073-2716-45c1-a38f-d8cc9cab01a3 tempest-ServersTestFqdnHostnames-1317897810 tempest-ServersTestFqdnHostnames-1317897810-project-member] Lock "3b852a19-9f86-486b-b98d-e87ffc54d943" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 162.974s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.182944] env[61243]: DEBUG nova.compute.manager [req-1da08f85-84d1-48c3-9e0c-4b73dec7376f req-d33f1a7f-8886-485c-a5f4-3cd4a27ffaec service nova] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Received event network-changed-8f597029-2d2c-4f15-8b33-593740cd82d6 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 747.183216] env[61243]: DEBUG nova.compute.manager [req-1da08f85-84d1-48c3-9e0c-4b73dec7376f req-d33f1a7f-8886-485c-a5f4-3cd4a27ffaec service nova] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Refreshing instance network info cache due to event network-changed-8f597029-2d2c-4f15-8b33-593740cd82d6. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 747.183429] env[61243]: DEBUG oslo_concurrency.lockutils [req-1da08f85-84d1-48c3-9e0c-4b73dec7376f req-d33f1a7f-8886-485c-a5f4-3cd4a27ffaec service nova] Acquiring lock "refresh_cache-61d68f1c-6a48-475a-8cf5-fa6ce3b09f20" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.183867] env[61243]: DEBUG oslo_concurrency.lockutils [req-1da08f85-84d1-48c3-9e0c-4b73dec7376f req-d33f1a7f-8886-485c-a5f4-3cd4a27ffaec service nova] Acquired lock "refresh_cache-61d68f1c-6a48-475a-8cf5-fa6ce3b09f20" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.183867] env[61243]: DEBUG nova.network.neutron [req-1da08f85-84d1-48c3-9e0c-4b73dec7376f req-d33f1a7f-8886-485c-a5f4-3cd4a27ffaec service nova] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Refreshing network info cache for port 8f597029-2d2c-4f15-8b33-593740cd82d6 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 747.228595] env[61243]: ERROR nova.compute.manager [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8f597029-2d2c-4f15-8b33-593740cd82d6, please check neutron logs for more information. [ 747.228595] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 747.228595] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.228595] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 747.228595] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.228595] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 747.228595] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.228595] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 747.228595] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.228595] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 747.228595] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.228595] env[61243]: ERROR nova.compute.manager raise self.value [ 747.228595] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.228595] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 747.228595] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.228595] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 747.229226] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.229226] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 747.229226] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8f597029-2d2c-4f15-8b33-593740cd82d6, please check neutron logs for more information. [ 747.229226] env[61243]: ERROR nova.compute.manager [ 747.229226] env[61243]: Traceback (most recent call last): [ 747.229226] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 747.229226] env[61243]: listener.cb(fileno) [ 747.229226] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.229226] env[61243]: result = function(*args, **kwargs) [ 747.229226] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 747.229226] env[61243]: return func(*args, **kwargs) [ 747.229226] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.229226] env[61243]: raise e [ 747.229226] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.229226] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 747.229226] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.229226] env[61243]: created_port_ids = self._update_ports_for_instance( [ 747.229226] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.229226] env[61243]: with excutils.save_and_reraise_exception(): [ 747.229226] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.229226] env[61243]: self.force_reraise() [ 747.229226] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.229226] env[61243]: raise self.value [ 747.229226] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.229226] env[61243]: updated_port = self._update_port( [ 747.229226] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.229226] env[61243]: _ensure_no_port_binding_failure(port) [ 747.229226] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.229226] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 747.230249] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 8f597029-2d2c-4f15-8b33-593740cd82d6, please check neutron logs for more information. [ 747.230249] env[61243]: Removing descriptor: 15 [ 747.377680] env[61243]: DEBUG oslo_concurrency.lockutils [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.811s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.380104] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.872s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.382962] env[61243]: DEBUG nova.compute.manager [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 747.409249] env[61243]: INFO nova.scheduler.client.report [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Deleted allocations for instance fd2d724a-76b6-405f-b003-54de2ad84f5a [ 747.588135] env[61243]: DEBUG nova.compute.manager [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 747.620151] env[61243]: DEBUG nova.virt.hardware [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 747.620151] env[61243]: DEBUG nova.virt.hardware [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 747.620151] env[61243]: DEBUG nova.virt.hardware [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 747.620330] env[61243]: DEBUG nova.virt.hardware [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 747.620330] env[61243]: DEBUG nova.virt.hardware [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 747.620330] env[61243]: DEBUG nova.virt.hardware [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 747.620616] env[61243]: DEBUG nova.virt.hardware [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 747.621140] env[61243]: DEBUG nova.virt.hardware [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 747.621563] env[61243]: DEBUG nova.virt.hardware [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 747.621913] env[61243]: DEBUG nova.virt.hardware [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 747.625017] env[61243]: DEBUG nova.virt.hardware [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 747.625017] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff815cab-368d-4260-9767-3abdfadd87bf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.636051] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d7784f5-8d42-42ad-aadf-1556c71e4ae3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.646936] env[61243]: ERROR nova.compute.manager [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8f597029-2d2c-4f15-8b33-593740cd82d6, please check neutron logs for more information. [ 747.646936] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Traceback (most recent call last): [ 747.646936] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 747.646936] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] yield resources [ 747.646936] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 747.646936] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] self.driver.spawn(context, instance, image_meta, [ 747.646936] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 747.646936] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] self._vmops.spawn(context, instance, image_meta, injected_files, [ 747.646936] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 747.646936] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] vm_ref = self.build_virtual_machine(instance, [ 747.646936] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 747.647430] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] vif_infos = vmwarevif.get_vif_info(self._session, [ 747.647430] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 747.647430] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] for vif in network_info: [ 747.647430] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 747.647430] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] return self._sync_wrapper(fn, *args, **kwargs) [ 747.647430] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 747.647430] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] self.wait() [ 747.647430] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 747.647430] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] self[:] = self._gt.wait() [ 747.647430] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 747.647430] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] return self._exit_event.wait() [ 747.647430] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 747.647430] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] current.throw(*self._exc) [ 747.647881] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.647881] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] result = function(*args, **kwargs) [ 747.647881] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 747.647881] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] return func(*args, **kwargs) [ 747.647881] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.647881] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] raise e [ 747.647881] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.647881] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] nwinfo = self.network_api.allocate_for_instance( [ 747.647881] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.647881] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] created_port_ids = self._update_ports_for_instance( [ 747.647881] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.647881] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] with excutils.save_and_reraise_exception(): [ 747.647881] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.648341] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] self.force_reraise() [ 747.648341] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.648341] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] raise self.value [ 747.648341] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.648341] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] updated_port = self._update_port( [ 747.648341] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.648341] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] _ensure_no_port_binding_failure(port) [ 747.648341] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.648341] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] raise exception.PortBindingFailed(port_id=port['id']) [ 747.648341] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] nova.exception.PortBindingFailed: Binding failed for port 8f597029-2d2c-4f15-8b33-593740cd82d6, please check neutron logs for more information. [ 747.648341] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] [ 747.651833] env[61243]: INFO nova.compute.manager [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Terminating instance [ 747.653180] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Acquiring lock "refresh_cache-61d68f1c-6a48-475a-8cf5-fa6ce3b09f20" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.703378] env[61243]: DEBUG nova.network.neutron [req-1da08f85-84d1-48c3-9e0c-4b73dec7376f req-d33f1a7f-8886-485c-a5f4-3cd4a27ffaec service nova] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.817651] env[61243]: DEBUG nova.network.neutron [req-1da08f85-84d1-48c3-9e0c-4b73dec7376f req-d33f1a7f-8886-485c-a5f4-3cd4a27ffaec service nova] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.906205] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.918337] env[61243]: DEBUG oslo_concurrency.lockutils [None req-65310ae5-655c-408e-965c-9ab8b4e72a52 tempest-ServersAaction247Test-990409856 tempest-ServersAaction247Test-990409856-project-member] Lock "fd2d724a-76b6-405f-b003-54de2ad84f5a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.978s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.132792] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed09fccb-a927-405d-a25f-eabdd1d841ad {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.140601] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd64ad5d-4f02-4030-b965-188b044e4ef1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.170615] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35fd0b50-2497-443f-bb09-2b586fdbe4b0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.178517] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-407bf9cd-1ad4-4851-8ac7-b5b843730b83 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.195064] env[61243]: DEBUG nova.compute.provider_tree [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.321364] env[61243]: DEBUG oslo_concurrency.lockutils [req-1da08f85-84d1-48c3-9e0c-4b73dec7376f req-d33f1a7f-8886-485c-a5f4-3cd4a27ffaec service nova] Releasing lock "refresh_cache-61d68f1c-6a48-475a-8cf5-fa6ce3b09f20" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.321364] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Acquired lock "refresh_cache-61d68f1c-6a48-475a-8cf5-fa6ce3b09f20" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.321364] env[61243]: DEBUG nova.network.neutron [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 748.697716] env[61243]: DEBUG nova.scheduler.client.report [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.846430] env[61243]: DEBUG nova.network.neutron [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.939623] env[61243]: DEBUG nova.network.neutron [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.202826] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.823s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.204055] env[61243]: ERROR nova.compute.manager [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1bf74449-72ce-4393-8d7d-821b435c9b32, please check neutron logs for more information. [ 749.204055] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] Traceback (most recent call last): [ 749.204055] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 749.204055] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] self.driver.spawn(context, instance, image_meta, [ 749.204055] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 749.204055] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.204055] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.204055] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] vm_ref = self.build_virtual_machine(instance, [ 749.204055] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.204055] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.204055] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.204467] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] for vif in network_info: [ 749.204467] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.204467] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] return self._sync_wrapper(fn, *args, **kwargs) [ 749.204467] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.204467] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] self.wait() [ 749.204467] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.204467] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] self[:] = self._gt.wait() [ 749.204467] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.204467] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] return self._exit_event.wait() [ 749.204467] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 749.204467] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] result = hub.switch() [ 749.204467] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 749.204467] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] return self.greenlet.switch() [ 749.204895] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.204895] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] result = function(*args, **kwargs) [ 749.204895] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 749.204895] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] return func(*args, **kwargs) [ 749.204895] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.204895] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] raise e [ 749.204895] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.204895] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] nwinfo = self.network_api.allocate_for_instance( [ 749.204895] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.204895] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] created_port_ids = self._update_ports_for_instance( [ 749.204895] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.204895] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] with excutils.save_and_reraise_exception(): [ 749.204895] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.205388] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] self.force_reraise() [ 749.205388] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.205388] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] raise self.value [ 749.205388] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.205388] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] updated_port = self._update_port( [ 749.205388] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.205388] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] _ensure_no_port_binding_failure(port) [ 749.205388] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.205388] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] raise exception.PortBindingFailed(port_id=port['id']) [ 749.205388] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] nova.exception.PortBindingFailed: Binding failed for port 1bf74449-72ce-4393-8d7d-821b435c9b32, please check neutron logs for more information. [ 749.205388] env[61243]: ERROR nova.compute.manager [instance: be8ac265-b214-4db5-9781-d171cee46538] [ 749.205806] env[61243]: DEBUG nova.compute.utils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Binding failed for port 1bf74449-72ce-4393-8d7d-821b435c9b32, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 749.205806] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.513s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.207730] env[61243]: INFO nova.compute.claims [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 749.211044] env[61243]: DEBUG nova.compute.manager [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Build of instance be8ac265-b214-4db5-9781-d171cee46538 was re-scheduled: Binding failed for port 1bf74449-72ce-4393-8d7d-821b435c9b32, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 749.211492] env[61243]: DEBUG nova.compute.manager [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 749.211720] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Acquiring lock "refresh_cache-be8ac265-b214-4db5-9781-d171cee46538" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.211861] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Acquired lock "refresh_cache-be8ac265-b214-4db5-9781-d171cee46538" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.212044] env[61243]: DEBUG nova.network.neutron [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 749.215157] env[61243]: DEBUG nova.compute.manager [req-adebbcac-82ba-4c8e-934b-1b788de25bfa req-18990694-9154-4afb-ae7c-17e8a5e0f1d0 service nova] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Received event network-vif-deleted-8f597029-2d2c-4f15-8b33-593740cd82d6 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.443436] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Releasing lock "refresh_cache-61d68f1c-6a48-475a-8cf5-fa6ce3b09f20" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.443572] env[61243]: DEBUG nova.compute.manager [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 749.443762] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 749.444088] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4b083071-d7ad-4bd4-b56f-84e36454655b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.453286] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b11278a-716c-4999-930f-46a2f8bacc5d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.478377] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20 could not be found. [ 749.478600] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 749.478783] env[61243]: INFO nova.compute.manager [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Took 0.04 seconds to destroy the instance on the hypervisor. [ 749.479119] env[61243]: DEBUG oslo.service.loopingcall [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.479316] env[61243]: DEBUG nova.compute.manager [-] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 749.479486] env[61243]: DEBUG nova.network.neutron [-] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 749.500210] env[61243]: DEBUG nova.network.neutron [-] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.739530] env[61243]: DEBUG nova.network.neutron [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.890437] env[61243]: DEBUG nova.network.neutron [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.001169] env[61243]: DEBUG nova.network.neutron [-] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.392372] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Releasing lock "refresh_cache-be8ac265-b214-4db5-9781-d171cee46538" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.392713] env[61243]: DEBUG nova.compute.manager [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 750.392871] env[61243]: DEBUG nova.compute.manager [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 750.392982] env[61243]: DEBUG nova.network.neutron [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 750.413949] env[61243]: DEBUG nova.network.neutron [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.504358] env[61243]: INFO nova.compute.manager [-] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Took 1.02 seconds to deallocate network for instance. [ 750.510462] env[61243]: DEBUG nova.compute.claims [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 750.510669] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.516815] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42bd459e-f08d-45dd-b595-f632a8bef200 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.525805] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7d407b-2f24-44a8-aebc-617378fc2f6a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.557755] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd91238-a2dd-4043-b6bd-202ee7409ef7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.565494] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb57b838-cb53-41de-9770-a1ce99ffc2cd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.579011] env[61243]: DEBUG nova.compute.provider_tree [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.922966] env[61243]: DEBUG nova.network.neutron [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.083095] env[61243]: DEBUG nova.scheduler.client.report [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.429695] env[61243]: INFO nova.compute.manager [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] [instance: be8ac265-b214-4db5-9781-d171cee46538] Took 1.04 seconds to deallocate network for instance. [ 751.588287] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.383s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.589352] env[61243]: DEBUG nova.compute.manager [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 751.593255] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.110s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.597233] env[61243]: INFO nova.compute.claims [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.105019] env[61243]: DEBUG nova.compute.utils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 752.105743] env[61243]: DEBUG nova.compute.manager [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 752.105743] env[61243]: DEBUG nova.network.neutron [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 752.154117] env[61243]: DEBUG nova.policy [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e30f3e88003426497d33d42274a8d07', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5c9a93490999434a9f1f152538ded354', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 752.427034] env[61243]: DEBUG nova.network.neutron [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Successfully created port: 2363167a-7627-41e4-9d85-beb565111760 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 752.473117] env[61243]: INFO nova.scheduler.client.report [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Deleted allocations for instance be8ac265-b214-4db5-9781-d171cee46538 [ 752.615750] env[61243]: DEBUG nova.compute.manager [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 752.908471] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a28108d1-02ca-4a87-a225-cf16a3448a0a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.917020] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7798c2fd-43ac-4ef4-b067-65f742019afd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.955882] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd29c18-4a73-4b4d-8d31-8598d313ee1f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.964521] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8256b9b-2c32-488c-9f97-1fe31b819d0c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.983414] env[61243]: DEBUG nova.compute.provider_tree [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.986670] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2903917-baa8-429a-ba0a-097482eb955e tempest-ServerMetadataTestJSON-1032477520 tempest-ServerMetadataTestJSON-1032477520-project-member] Lock "be8ac265-b214-4db5-9781-d171cee46538" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 165.840s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.050484] env[61243]: DEBUG oslo_concurrency.lockutils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "894eb798-6c9e-47cb-8eb5-4610fe184bc5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.050484] env[61243]: DEBUG oslo_concurrency.lockutils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "894eb798-6c9e-47cb-8eb5-4610fe184bc5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.290301] env[61243]: DEBUG nova.compute.manager [req-a2a2565f-5058-442d-89ba-569fd91ea061 req-9d3a48a9-c11d-4961-9e51-863cb47300ed service nova] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Received event network-changed-2363167a-7627-41e4-9d85-beb565111760 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 753.291411] env[61243]: DEBUG nova.compute.manager [req-a2a2565f-5058-442d-89ba-569fd91ea061 req-9d3a48a9-c11d-4961-9e51-863cb47300ed service nova] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Refreshing instance network info cache due to event network-changed-2363167a-7627-41e4-9d85-beb565111760. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 753.292146] env[61243]: DEBUG oslo_concurrency.lockutils [req-a2a2565f-5058-442d-89ba-569fd91ea061 req-9d3a48a9-c11d-4961-9e51-863cb47300ed service nova] Acquiring lock "refresh_cache-454d97a2-40d0-4c7a-9df7-04c391715e0d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.292328] env[61243]: DEBUG oslo_concurrency.lockutils [req-a2a2565f-5058-442d-89ba-569fd91ea061 req-9d3a48a9-c11d-4961-9e51-863cb47300ed service nova] Acquired lock "refresh_cache-454d97a2-40d0-4c7a-9df7-04c391715e0d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.292499] env[61243]: DEBUG nova.network.neutron [req-a2a2565f-5058-442d-89ba-569fd91ea061 req-9d3a48a9-c11d-4961-9e51-863cb47300ed service nova] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Refreshing network info cache for port 2363167a-7627-41e4-9d85-beb565111760 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 753.418662] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.419866] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.444591] env[61243]: ERROR nova.compute.manager [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2363167a-7627-41e4-9d85-beb565111760, please check neutron logs for more information. [ 753.444591] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 753.444591] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.444591] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 753.444591] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.444591] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 753.444591] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.444591] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 753.444591] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.444591] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 753.444591] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.444591] env[61243]: ERROR nova.compute.manager raise self.value [ 753.444591] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.444591] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 753.444591] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.444591] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 753.445182] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.445182] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 753.445182] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2363167a-7627-41e4-9d85-beb565111760, please check neutron logs for more information. [ 753.445182] env[61243]: ERROR nova.compute.manager [ 753.445182] env[61243]: Traceback (most recent call last): [ 753.445182] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 753.445182] env[61243]: listener.cb(fileno) [ 753.445182] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.445182] env[61243]: result = function(*args, **kwargs) [ 753.445182] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 753.445182] env[61243]: return func(*args, **kwargs) [ 753.445182] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.445182] env[61243]: raise e [ 753.445182] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.445182] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 753.445182] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.445182] env[61243]: created_port_ids = self._update_ports_for_instance( [ 753.445182] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.445182] env[61243]: with excutils.save_and_reraise_exception(): [ 753.445182] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.445182] env[61243]: self.force_reraise() [ 753.445182] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.445182] env[61243]: raise self.value [ 753.445182] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.445182] env[61243]: updated_port = self._update_port( [ 753.445182] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.445182] env[61243]: _ensure_no_port_binding_failure(port) [ 753.445182] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.445182] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 753.446756] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 2363167a-7627-41e4-9d85-beb565111760, please check neutron logs for more information. [ 753.446756] env[61243]: Removing descriptor: 15 [ 753.487514] env[61243]: DEBUG nova.scheduler.client.report [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 753.490858] env[61243]: DEBUG nova.compute.manager [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 753.628774] env[61243]: DEBUG nova.compute.manager [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 753.662897] env[61243]: DEBUG nova.virt.hardware [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 753.663179] env[61243]: DEBUG nova.virt.hardware [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 753.663339] env[61243]: DEBUG nova.virt.hardware [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 753.663521] env[61243]: DEBUG nova.virt.hardware [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 753.663667] env[61243]: DEBUG nova.virt.hardware [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 753.663853] env[61243]: DEBUG nova.virt.hardware [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 753.664372] env[61243]: DEBUG nova.virt.hardware [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 753.664646] env[61243]: DEBUG nova.virt.hardware [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 753.664890] env[61243]: DEBUG nova.virt.hardware [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 753.665446] env[61243]: DEBUG nova.virt.hardware [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 753.665700] env[61243]: DEBUG nova.virt.hardware [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 753.666804] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d91fb65-4a83-4dbd-b530-d40c54d148d6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.677331] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf883ab-1f89-4027-b38a-73b2f50fb35e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.693872] env[61243]: ERROR nova.compute.manager [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2363167a-7627-41e4-9d85-beb565111760, please check neutron logs for more information. [ 753.693872] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Traceback (most recent call last): [ 753.693872] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 753.693872] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] yield resources [ 753.693872] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 753.693872] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] self.driver.spawn(context, instance, image_meta, [ 753.693872] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 753.693872] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.693872] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.693872] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] vm_ref = self.build_virtual_machine(instance, [ 753.693872] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.694242] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.694242] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.694242] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] for vif in network_info: [ 753.694242] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.694242] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] return self._sync_wrapper(fn, *args, **kwargs) [ 753.694242] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.694242] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] self.wait() [ 753.694242] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.694242] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] self[:] = self._gt.wait() [ 753.694242] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.694242] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] return self._exit_event.wait() [ 753.694242] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 753.694242] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] current.throw(*self._exc) [ 753.694787] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.694787] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] result = function(*args, **kwargs) [ 753.694787] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 753.694787] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] return func(*args, **kwargs) [ 753.694787] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.694787] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] raise e [ 753.694787] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.694787] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] nwinfo = self.network_api.allocate_for_instance( [ 753.694787] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.694787] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] created_port_ids = self._update_ports_for_instance( [ 753.694787] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.694787] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] with excutils.save_and_reraise_exception(): [ 753.694787] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.695176] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] self.force_reraise() [ 753.695176] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.695176] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] raise self.value [ 753.695176] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.695176] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] updated_port = self._update_port( [ 753.695176] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.695176] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] _ensure_no_port_binding_failure(port) [ 753.695176] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.695176] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] raise exception.PortBindingFailed(port_id=port['id']) [ 753.695176] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] nova.exception.PortBindingFailed: Binding failed for port 2363167a-7627-41e4-9d85-beb565111760, please check neutron logs for more information. [ 753.695176] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] [ 753.695176] env[61243]: INFO nova.compute.manager [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Terminating instance [ 753.697705] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Acquiring lock "refresh_cache-454d97a2-40d0-4c7a-9df7-04c391715e0d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.813228] env[61243]: DEBUG nova.network.neutron [req-a2a2565f-5058-442d-89ba-569fd91ea061 req-9d3a48a9-c11d-4961-9e51-863cb47300ed service nova] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.897775] env[61243]: DEBUG nova.network.neutron [req-a2a2565f-5058-442d-89ba-569fd91ea061 req-9d3a48a9-c11d-4961-9e51-863cb47300ed service nova] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.998113] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.998113] env[61243]: DEBUG nova.compute.manager [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 754.000968] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.462s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.002632] env[61243]: INFO nova.compute.claims [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 754.023765] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.403227] env[61243]: DEBUG oslo_concurrency.lockutils [req-a2a2565f-5058-442d-89ba-569fd91ea061 req-9d3a48a9-c11d-4961-9e51-863cb47300ed service nova] Releasing lock "refresh_cache-454d97a2-40d0-4c7a-9df7-04c391715e0d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.403794] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Acquired lock "refresh_cache-454d97a2-40d0-4c7a-9df7-04c391715e0d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.404102] env[61243]: DEBUG nova.network.neutron [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 754.509119] env[61243]: DEBUG nova.compute.utils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 754.516533] env[61243]: DEBUG nova.compute.manager [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 754.517220] env[61243]: DEBUG nova.network.neutron [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 754.579619] env[61243]: DEBUG nova.policy [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3eb1349f8af545dfb8bf9d0e21803204', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfc64d5461ab4b2aa19a87c197ac8816', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 754.939511] env[61243]: DEBUG nova.network.neutron [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.017631] env[61243]: DEBUG nova.compute.manager [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 755.081627] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 755.081823] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 755.090413] env[61243]: DEBUG nova.network.neutron [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.091985] env[61243]: DEBUG nova.network.neutron [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Successfully created port: f6efca39-d061-4332-a8c2-76d8cfc414b3 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 755.365217] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25998066-9c75-4bca-8383-0b908e946e86 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.374674] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-043f4122-938a-49b7-a12e-46342ff38e44 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.407251] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f17b5bcc-9323-48d4-852e-3c53488122b9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.419734] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9527251-d8d2-41ef-8a1c-6ff940b0459e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.426218] env[61243]: DEBUG nova.compute.manager [req-29bf68d4-227a-4170-822e-d53421c83a40 req-161ffaee-8c25-4108-b1b4-3fde57479313 service nova] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Received event network-vif-deleted-2363167a-7627-41e4-9d85-beb565111760 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 755.438672] env[61243]: DEBUG nova.compute.provider_tree [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.589413] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 755.589413] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Starting heal instance info cache {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 755.589413] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Rebuilding the list of instances to heal {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 755.598090] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Releasing lock "refresh_cache-454d97a2-40d0-4c7a-9df7-04c391715e0d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.598090] env[61243]: DEBUG nova.compute.manager [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 755.598090] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 755.598090] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5e42d2b9-c05b-4985-960d-f92c9ebb2f7b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.614989] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7378e32-769a-4158-8d5e-71e11c07187b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.649789] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 454d97a2-40d0-4c7a-9df7-04c391715e0d could not be found. [ 755.650430] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 755.650876] env[61243]: INFO nova.compute.manager [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Took 0.06 seconds to destroy the instance on the hypervisor. [ 755.651408] env[61243]: DEBUG oslo.service.loopingcall [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 755.651925] env[61243]: DEBUG nova.compute.manager [-] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 755.652233] env[61243]: DEBUG nova.network.neutron [-] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 755.676129] env[61243]: DEBUG nova.network.neutron [-] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.943300] env[61243]: DEBUG nova.scheduler.client.report [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 756.033752] env[61243]: DEBUG nova.compute.manager [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 756.065748] env[61243]: DEBUG nova.virt.hardware [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 756.065994] env[61243]: DEBUG nova.virt.hardware [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 756.066167] env[61243]: DEBUG nova.virt.hardware [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 756.066346] env[61243]: DEBUG nova.virt.hardware [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 756.066490] env[61243]: DEBUG nova.virt.hardware [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 756.066632] env[61243]: DEBUG nova.virt.hardware [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 756.066831] env[61243]: DEBUG nova.virt.hardware [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 756.066991] env[61243]: DEBUG nova.virt.hardware [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 756.067169] env[61243]: DEBUG nova.virt.hardware [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 756.067332] env[61243]: DEBUG nova.virt.hardware [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 756.067502] env[61243]: DEBUG nova.virt.hardware [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 756.068399] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ff9fc7-2469-4672-a68d-c403a61681fd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.076599] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d3e378-b1c9-4f12-92bf-3a739d3480bf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.092878] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Skipping network cache update for instance because it is Building. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 756.092878] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Skipping network cache update for instance because it is Building. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 756.093039] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Skipping network cache update for instance because it is Building. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 756.093150] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Skipping network cache update for instance because it is Building. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 756.093275] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Skipping network cache update for instance because it is Building. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 756.093393] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Skipping network cache update for instance because it is Building. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 756.093511] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Didn't find any instances for network info cache update. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 756.094027] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 756.094194] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 756.094347] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 756.094500] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 756.094658] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 756.094787] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 756.095082] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61243) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 756.095082] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager.update_available_resource {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 756.176659] env[61243]: DEBUG nova.network.neutron [-] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.225780] env[61243]: ERROR nova.compute.manager [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f6efca39-d061-4332-a8c2-76d8cfc414b3, please check neutron logs for more information. [ 756.225780] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 756.225780] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.225780] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 756.225780] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.225780] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 756.225780] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.225780] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 756.225780] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.225780] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 756.225780] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.225780] env[61243]: ERROR nova.compute.manager raise self.value [ 756.225780] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.225780] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 756.225780] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.225780] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 756.226401] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.226401] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 756.226401] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f6efca39-d061-4332-a8c2-76d8cfc414b3, please check neutron logs for more information. [ 756.226401] env[61243]: ERROR nova.compute.manager [ 756.226401] env[61243]: Traceback (most recent call last): [ 756.226401] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 756.226401] env[61243]: listener.cb(fileno) [ 756.226401] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.226401] env[61243]: result = function(*args, **kwargs) [ 756.226401] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 756.226401] env[61243]: return func(*args, **kwargs) [ 756.226401] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.226401] env[61243]: raise e [ 756.226401] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.226401] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 756.226401] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.226401] env[61243]: created_port_ids = self._update_ports_for_instance( [ 756.226401] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.226401] env[61243]: with excutils.save_and_reraise_exception(): [ 756.226401] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.226401] env[61243]: self.force_reraise() [ 756.226401] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.226401] env[61243]: raise self.value [ 756.226401] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.226401] env[61243]: updated_port = self._update_port( [ 756.226401] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.226401] env[61243]: _ensure_no_port_binding_failure(port) [ 756.226401] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.226401] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 756.227195] env[61243]: nova.exception.PortBindingFailed: Binding failed for port f6efca39-d061-4332-a8c2-76d8cfc414b3, please check neutron logs for more information. [ 756.227195] env[61243]: Removing descriptor: 15 [ 756.227195] env[61243]: ERROR nova.compute.manager [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f6efca39-d061-4332-a8c2-76d8cfc414b3, please check neutron logs for more information. [ 756.227195] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Traceback (most recent call last): [ 756.227195] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 756.227195] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] yield resources [ 756.227195] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 756.227195] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] self.driver.spawn(context, instance, image_meta, [ 756.227195] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 756.227195] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 756.227195] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 756.227195] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] vm_ref = self.build_virtual_machine(instance, [ 756.227650] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 756.227650] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] vif_infos = vmwarevif.get_vif_info(self._session, [ 756.227650] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 756.227650] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] for vif in network_info: [ 756.227650] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 756.227650] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] return self._sync_wrapper(fn, *args, **kwargs) [ 756.227650] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 756.227650] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] self.wait() [ 756.227650] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 756.227650] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] self[:] = self._gt.wait() [ 756.227650] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 756.227650] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] return self._exit_event.wait() [ 756.227650] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 756.228135] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] result = hub.switch() [ 756.228135] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 756.228135] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] return self.greenlet.switch() [ 756.228135] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.228135] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] result = function(*args, **kwargs) [ 756.228135] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 756.228135] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] return func(*args, **kwargs) [ 756.228135] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.228135] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] raise e [ 756.228135] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.228135] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] nwinfo = self.network_api.allocate_for_instance( [ 756.228135] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.228135] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] created_port_ids = self._update_ports_for_instance( [ 756.228621] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.228621] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] with excutils.save_and_reraise_exception(): [ 756.228621] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.228621] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] self.force_reraise() [ 756.228621] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.228621] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] raise self.value [ 756.228621] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.228621] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] updated_port = self._update_port( [ 756.228621] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.228621] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] _ensure_no_port_binding_failure(port) [ 756.228621] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.228621] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] raise exception.PortBindingFailed(port_id=port['id']) [ 756.228998] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] nova.exception.PortBindingFailed: Binding failed for port f6efca39-d061-4332-a8c2-76d8cfc414b3, please check neutron logs for more information. [ 756.228998] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] [ 756.228998] env[61243]: INFO nova.compute.manager [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Terminating instance [ 756.229117] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Acquiring lock "refresh_cache-2e25cb9a-9b21-4b60-b741-944e53bd50e9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.229182] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Acquired lock "refresh_cache-2e25cb9a-9b21-4b60-b741-944e53bd50e9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.229347] env[61243]: DEBUG nova.network.neutron [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 756.453023] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.453023] env[61243]: DEBUG nova.compute.manager [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 756.455618] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.041s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.458330] env[61243]: INFO nova.compute.claims [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.597850] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.680100] env[61243]: INFO nova.compute.manager [-] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Took 1.03 seconds to deallocate network for instance. [ 756.682325] env[61243]: DEBUG nova.compute.claims [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 756.682503] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.749123] env[61243]: DEBUG nova.network.neutron [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.822906] env[61243]: DEBUG nova.network.neutron [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.962790] env[61243]: DEBUG nova.compute.utils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 756.966224] env[61243]: DEBUG nova.compute.manager [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 756.966402] env[61243]: DEBUG nova.network.neutron [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 757.003150] env[61243]: DEBUG nova.policy [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a9e6dc85a31d4ec38efe0c750c49ead1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '113713170fb94b8a8d7985b9a02faf77', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 757.247170] env[61243]: DEBUG nova.network.neutron [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Successfully created port: dfcaaed7-6129-4957-b854-e8213e476069 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 757.326732] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Releasing lock "refresh_cache-2e25cb9a-9b21-4b60-b741-944e53bd50e9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.327285] env[61243]: DEBUG nova.compute.manager [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 757.327554] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 757.327908] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-01b8a51a-77a1-44fc-bb7a-db67aee349d9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.339941] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5899752f-6631-4aea-83d3-d811d86d093e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.366029] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2e25cb9a-9b21-4b60-b741-944e53bd50e9 could not be found. [ 757.366029] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 757.366029] env[61243]: INFO nova.compute.manager [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 757.366029] env[61243]: DEBUG oslo.service.loopingcall [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 757.366029] env[61243]: DEBUG nova.compute.manager [-] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 757.366029] env[61243]: DEBUG nova.network.neutron [-] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 757.383265] env[61243]: DEBUG nova.network.neutron [-] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.454687] env[61243]: DEBUG nova.compute.manager [req-f0bbecb1-3736-4072-8758-23a7d06b54c1 req-71d21589-7202-45ee-bff2-ed15dfa8dec2 service nova] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Received event network-changed-f6efca39-d061-4332-a8c2-76d8cfc414b3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 757.454891] env[61243]: DEBUG nova.compute.manager [req-f0bbecb1-3736-4072-8758-23a7d06b54c1 req-71d21589-7202-45ee-bff2-ed15dfa8dec2 service nova] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Refreshing instance network info cache due to event network-changed-f6efca39-d061-4332-a8c2-76d8cfc414b3. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 757.459209] env[61243]: DEBUG oslo_concurrency.lockutils [req-f0bbecb1-3736-4072-8758-23a7d06b54c1 req-71d21589-7202-45ee-bff2-ed15dfa8dec2 service nova] Acquiring lock "refresh_cache-2e25cb9a-9b21-4b60-b741-944e53bd50e9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.459415] env[61243]: DEBUG oslo_concurrency.lockutils [req-f0bbecb1-3736-4072-8758-23a7d06b54c1 req-71d21589-7202-45ee-bff2-ed15dfa8dec2 service nova] Acquired lock "refresh_cache-2e25cb9a-9b21-4b60-b741-944e53bd50e9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.459590] env[61243]: DEBUG nova.network.neutron [req-f0bbecb1-3736-4072-8758-23a7d06b54c1 req-71d21589-7202-45ee-bff2-ed15dfa8dec2 service nova] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Refreshing network info cache for port f6efca39-d061-4332-a8c2-76d8cfc414b3 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 757.467610] env[61243]: DEBUG nova.compute.manager [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 757.752718] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36fb58ea-5495-4cf7-a46e-b506fda951ed {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.761521] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577f16d4-d179-47f6-b79e-7febbd70e0dd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.794830] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4fb19c-2a8a-459c-a525-b841039ef579 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.803337] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed20a409-dc2b-4bd2-b4c3-64481c8552d4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.819993] env[61243]: DEBUG nova.compute.provider_tree [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.891042] env[61243]: DEBUG nova.network.neutron [-] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.988501] env[61243]: DEBUG nova.network.neutron [req-f0bbecb1-3736-4072-8758-23a7d06b54c1 req-71d21589-7202-45ee-bff2-ed15dfa8dec2 service nova] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.113847] env[61243]: DEBUG nova.network.neutron [req-f0bbecb1-3736-4072-8758-23a7d06b54c1 req-71d21589-7202-45ee-bff2-ed15dfa8dec2 service nova] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.208686] env[61243]: ERROR nova.compute.manager [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dfcaaed7-6129-4957-b854-e8213e476069, please check neutron logs for more information. [ 758.208686] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 758.208686] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.208686] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 758.208686] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.208686] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 758.208686] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.208686] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 758.208686] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.208686] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 758.208686] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.208686] env[61243]: ERROR nova.compute.manager raise self.value [ 758.208686] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.208686] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 758.208686] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.208686] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 758.209454] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.209454] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 758.209454] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dfcaaed7-6129-4957-b854-e8213e476069, please check neutron logs for more information. [ 758.209454] env[61243]: ERROR nova.compute.manager [ 758.209454] env[61243]: Traceback (most recent call last): [ 758.209454] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 758.209454] env[61243]: listener.cb(fileno) [ 758.209454] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.209454] env[61243]: result = function(*args, **kwargs) [ 758.209454] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 758.209454] env[61243]: return func(*args, **kwargs) [ 758.209454] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.209454] env[61243]: raise e [ 758.209454] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.209454] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 758.209454] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.209454] env[61243]: created_port_ids = self._update_ports_for_instance( [ 758.209454] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.209454] env[61243]: with excutils.save_and_reraise_exception(): [ 758.209454] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.209454] env[61243]: self.force_reraise() [ 758.209454] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.209454] env[61243]: raise self.value [ 758.209454] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.209454] env[61243]: updated_port = self._update_port( [ 758.209454] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.209454] env[61243]: _ensure_no_port_binding_failure(port) [ 758.209454] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.209454] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 758.210613] env[61243]: nova.exception.PortBindingFailed: Binding failed for port dfcaaed7-6129-4957-b854-e8213e476069, please check neutron logs for more information. [ 758.210613] env[61243]: Removing descriptor: 15 [ 758.323713] env[61243]: DEBUG nova.scheduler.client.report [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 758.394415] env[61243]: INFO nova.compute.manager [-] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Took 1.03 seconds to deallocate network for instance. [ 758.396928] env[61243]: DEBUG nova.compute.claims [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 758.396928] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.481217] env[61243]: DEBUG nova.compute.manager [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 758.507292] env[61243]: DEBUG nova.virt.hardware [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 758.507548] env[61243]: DEBUG nova.virt.hardware [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 758.507706] env[61243]: DEBUG nova.virt.hardware [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 758.507890] env[61243]: DEBUG nova.virt.hardware [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 758.508050] env[61243]: DEBUG nova.virt.hardware [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 758.508203] env[61243]: DEBUG nova.virt.hardware [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 758.508409] env[61243]: DEBUG nova.virt.hardware [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 758.508571] env[61243]: DEBUG nova.virt.hardware [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 758.508734] env[61243]: DEBUG nova.virt.hardware [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 758.508896] env[61243]: DEBUG nova.virt.hardware [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 758.509090] env[61243]: DEBUG nova.virt.hardware [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 758.509937] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07a64e5-702f-4a7a-8324-3abab718c19b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.518321] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc1862ef-995b-42a5-8660-9c573911b39a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.531935] env[61243]: ERROR nova.compute.manager [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dfcaaed7-6129-4957-b854-e8213e476069, please check neutron logs for more information. [ 758.531935] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Traceback (most recent call last): [ 758.531935] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 758.531935] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] yield resources [ 758.531935] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 758.531935] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] self.driver.spawn(context, instance, image_meta, [ 758.531935] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 758.531935] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] self._vmops.spawn(context, instance, image_meta, injected_files, [ 758.531935] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 758.531935] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] vm_ref = self.build_virtual_machine(instance, [ 758.531935] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 758.532427] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] vif_infos = vmwarevif.get_vif_info(self._session, [ 758.532427] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 758.532427] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] for vif in network_info: [ 758.532427] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 758.532427] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] return self._sync_wrapper(fn, *args, **kwargs) [ 758.532427] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 758.532427] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] self.wait() [ 758.532427] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 758.532427] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] self[:] = self._gt.wait() [ 758.532427] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 758.532427] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] return self._exit_event.wait() [ 758.532427] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 758.532427] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] current.throw(*self._exc) [ 758.533022] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.533022] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] result = function(*args, **kwargs) [ 758.533022] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 758.533022] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] return func(*args, **kwargs) [ 758.533022] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.533022] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] raise e [ 758.533022] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.533022] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] nwinfo = self.network_api.allocate_for_instance( [ 758.533022] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.533022] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] created_port_ids = self._update_ports_for_instance( [ 758.533022] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.533022] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] with excutils.save_and_reraise_exception(): [ 758.533022] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.533478] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] self.force_reraise() [ 758.533478] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.533478] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] raise self.value [ 758.533478] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.533478] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] updated_port = self._update_port( [ 758.533478] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.533478] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] _ensure_no_port_binding_failure(port) [ 758.533478] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.533478] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] raise exception.PortBindingFailed(port_id=port['id']) [ 758.533478] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] nova.exception.PortBindingFailed: Binding failed for port dfcaaed7-6129-4957-b854-e8213e476069, please check neutron logs for more information. [ 758.533478] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] [ 758.533478] env[61243]: INFO nova.compute.manager [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Terminating instance [ 758.534620] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "refresh_cache-4a0a08ca-bbf5-4124-853e-0be45f531d91" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.534781] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "refresh_cache-4a0a08ca-bbf5-4124-853e-0be45f531d91" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.534943] env[61243]: DEBUG nova.network.neutron [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 758.615789] env[61243]: DEBUG oslo_concurrency.lockutils [req-f0bbecb1-3736-4072-8758-23a7d06b54c1 req-71d21589-7202-45ee-bff2-ed15dfa8dec2 service nova] Releasing lock "refresh_cache-2e25cb9a-9b21-4b60-b741-944e53bd50e9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.615789] env[61243]: DEBUG nova.compute.manager [req-f0bbecb1-3736-4072-8758-23a7d06b54c1 req-71d21589-7202-45ee-bff2-ed15dfa8dec2 service nova] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Received event network-vif-deleted-f6efca39-d061-4332-a8c2-76d8cfc414b3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 758.828905] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.829442] env[61243]: DEBUG nova.compute.manager [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 758.832310] env[61243]: DEBUG oslo_concurrency.lockutils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.563s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.834523] env[61243]: INFO nova.compute.claims [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 759.055572] env[61243]: DEBUG nova.network.neutron [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.134246] env[61243]: DEBUG nova.network.neutron [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.338804] env[61243]: DEBUG nova.compute.utils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 759.342802] env[61243]: DEBUG nova.compute.manager [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 759.343042] env[61243]: DEBUG nova.network.neutron [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 759.380661] env[61243]: DEBUG nova.policy [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0127d3b30d75463fa35519795369f272', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5190be93eab049ad9b4e06bc257fe0ae', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 759.483863] env[61243]: DEBUG nova.compute.manager [req-bbe1a51f-eedb-40c4-8aff-e4632b69e13c req-a24a053c-570e-4304-8e73-99308759eeae service nova] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Received event network-changed-dfcaaed7-6129-4957-b854-e8213e476069 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.483863] env[61243]: DEBUG nova.compute.manager [req-bbe1a51f-eedb-40c4-8aff-e4632b69e13c req-a24a053c-570e-4304-8e73-99308759eeae service nova] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Refreshing instance network info cache due to event network-changed-dfcaaed7-6129-4957-b854-e8213e476069. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 759.483863] env[61243]: DEBUG oslo_concurrency.lockutils [req-bbe1a51f-eedb-40c4-8aff-e4632b69e13c req-a24a053c-570e-4304-8e73-99308759eeae service nova] Acquiring lock "refresh_cache-4a0a08ca-bbf5-4124-853e-0be45f531d91" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.640023] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "refresh_cache-4a0a08ca-bbf5-4124-853e-0be45f531d91" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.640023] env[61243]: DEBUG nova.compute.manager [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 759.640023] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 759.640023] env[61243]: DEBUG oslo_concurrency.lockutils [req-bbe1a51f-eedb-40c4-8aff-e4632b69e13c req-a24a053c-570e-4304-8e73-99308759eeae service nova] Acquired lock "refresh_cache-4a0a08ca-bbf5-4124-853e-0be45f531d91" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.640023] env[61243]: DEBUG nova.network.neutron [req-bbe1a51f-eedb-40c4-8aff-e4632b69e13c req-a24a053c-570e-4304-8e73-99308759eeae service nova] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Refreshing network info cache for port dfcaaed7-6129-4957-b854-e8213e476069 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 759.641435] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-144686c6-65d8-4158-bb8e-4615e9bd124f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.651295] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc2a2f9-7377-46be-91d5-961868a824a5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.664218] env[61243]: DEBUG nova.network.neutron [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Successfully created port: 708bace8-628b-46fb-a8c8-4e9364b1824e {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 759.678098] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4a0a08ca-bbf5-4124-853e-0be45f531d91 could not be found. [ 759.678098] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 759.678098] env[61243]: INFO nova.compute.manager [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Took 0.04 seconds to destroy the instance on the hypervisor. [ 759.678098] env[61243]: DEBUG oslo.service.loopingcall [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.678335] env[61243]: DEBUG nova.compute.manager [-] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.678335] env[61243]: DEBUG nova.network.neutron [-] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 759.693711] env[61243]: DEBUG nova.network.neutron [-] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.845827] env[61243]: DEBUG nova.compute.manager [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 760.127465] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3540a51-f0db-47df-abb2-43d3e39897f8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.135473] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a46f393-d1b7-48df-97dd-a9eabc06010c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.166295] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d5d9558-efe2-4c96-bc6e-8b9f476db7ae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.173305] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e0558d-5be9-44d4-ab15-45e076934424 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.187274] env[61243]: DEBUG nova.compute.provider_tree [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.192143] env[61243]: DEBUG nova.network.neutron [req-bbe1a51f-eedb-40c4-8aff-e4632b69e13c req-a24a053c-570e-4304-8e73-99308759eeae service nova] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.195953] env[61243]: DEBUG nova.network.neutron [-] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.300812] env[61243]: DEBUG nova.network.neutron [req-bbe1a51f-eedb-40c4-8aff-e4632b69e13c req-a24a053c-570e-4304-8e73-99308759eeae service nova] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.592517] env[61243]: ERROR nova.compute.manager [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 708bace8-628b-46fb-a8c8-4e9364b1824e, please check neutron logs for more information. [ 760.592517] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 760.592517] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.592517] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 760.592517] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.592517] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 760.592517] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.592517] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 760.592517] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.592517] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 760.592517] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.592517] env[61243]: ERROR nova.compute.manager raise self.value [ 760.592517] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.592517] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 760.592517] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.592517] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 760.593114] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.593114] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 760.593114] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 708bace8-628b-46fb-a8c8-4e9364b1824e, please check neutron logs for more information. [ 760.593114] env[61243]: ERROR nova.compute.manager [ 760.593114] env[61243]: Traceback (most recent call last): [ 760.593114] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 760.593114] env[61243]: listener.cb(fileno) [ 760.593114] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.593114] env[61243]: result = function(*args, **kwargs) [ 760.593114] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 760.593114] env[61243]: return func(*args, **kwargs) [ 760.593114] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.593114] env[61243]: raise e [ 760.593114] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.593114] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 760.593114] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.593114] env[61243]: created_port_ids = self._update_ports_for_instance( [ 760.593114] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.593114] env[61243]: with excutils.save_and_reraise_exception(): [ 760.593114] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.593114] env[61243]: self.force_reraise() [ 760.593114] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.593114] env[61243]: raise self.value [ 760.593114] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.593114] env[61243]: updated_port = self._update_port( [ 760.593114] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.593114] env[61243]: _ensure_no_port_binding_failure(port) [ 760.593114] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.593114] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 760.594494] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 708bace8-628b-46fb-a8c8-4e9364b1824e, please check neutron logs for more information. [ 760.594494] env[61243]: Removing descriptor: 15 [ 760.691061] env[61243]: DEBUG nova.scheduler.client.report [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.698087] env[61243]: INFO nova.compute.manager [-] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Took 1.02 seconds to deallocate network for instance. [ 760.700111] env[61243]: DEBUG nova.compute.claims [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 760.700299] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.803614] env[61243]: DEBUG oslo_concurrency.lockutils [req-bbe1a51f-eedb-40c4-8aff-e4632b69e13c req-a24a053c-570e-4304-8e73-99308759eeae service nova] Releasing lock "refresh_cache-4a0a08ca-bbf5-4124-853e-0be45f531d91" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.803877] env[61243]: DEBUG nova.compute.manager [req-bbe1a51f-eedb-40c4-8aff-e4632b69e13c req-a24a053c-570e-4304-8e73-99308759eeae service nova] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Received event network-vif-deleted-dfcaaed7-6129-4957-b854-e8213e476069 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 760.855979] env[61243]: DEBUG nova.compute.manager [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 760.880121] env[61243]: DEBUG nova.virt.hardware [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 760.880447] env[61243]: DEBUG nova.virt.hardware [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 760.880673] env[61243]: DEBUG nova.virt.hardware [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 760.880890] env[61243]: DEBUG nova.virt.hardware [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 760.881054] env[61243]: DEBUG nova.virt.hardware [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 760.881207] env[61243]: DEBUG nova.virt.hardware [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 760.881417] env[61243]: DEBUG nova.virt.hardware [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 760.881573] env[61243]: DEBUG nova.virt.hardware [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 760.881733] env[61243]: DEBUG nova.virt.hardware [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 760.881891] env[61243]: DEBUG nova.virt.hardware [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 760.882073] env[61243]: DEBUG nova.virt.hardware [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 760.883144] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a681414b-c10a-4d10-ac09-5de1e62cf35b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.891056] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f0272ae-9806-4bd2-a2e5-f1ce6d3a1954 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.905223] env[61243]: ERROR nova.compute.manager [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 708bace8-628b-46fb-a8c8-4e9364b1824e, please check neutron logs for more information. [ 760.905223] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Traceback (most recent call last): [ 760.905223] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 760.905223] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] yield resources [ 760.905223] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 760.905223] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] self.driver.spawn(context, instance, image_meta, [ 760.905223] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 760.905223] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 760.905223] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 760.905223] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] vm_ref = self.build_virtual_machine(instance, [ 760.905223] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 760.905818] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] vif_infos = vmwarevif.get_vif_info(self._session, [ 760.905818] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 760.905818] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] for vif in network_info: [ 760.905818] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 760.905818] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] return self._sync_wrapper(fn, *args, **kwargs) [ 760.905818] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 760.905818] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] self.wait() [ 760.905818] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 760.905818] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] self[:] = self._gt.wait() [ 760.905818] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 760.905818] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] return self._exit_event.wait() [ 760.905818] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 760.905818] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] current.throw(*self._exc) [ 760.906247] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.906247] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] result = function(*args, **kwargs) [ 760.906247] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 760.906247] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] return func(*args, **kwargs) [ 760.906247] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.906247] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] raise e [ 760.906247] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.906247] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] nwinfo = self.network_api.allocate_for_instance( [ 760.906247] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.906247] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] created_port_ids = self._update_ports_for_instance( [ 760.906247] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.906247] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] with excutils.save_and_reraise_exception(): [ 760.906247] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.906652] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] self.force_reraise() [ 760.906652] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.906652] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] raise self.value [ 760.906652] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.906652] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] updated_port = self._update_port( [ 760.906652] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.906652] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] _ensure_no_port_binding_failure(port) [ 760.906652] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.906652] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] raise exception.PortBindingFailed(port_id=port['id']) [ 760.906652] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] nova.exception.PortBindingFailed: Binding failed for port 708bace8-628b-46fb-a8c8-4e9364b1824e, please check neutron logs for more information. [ 760.906652] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] [ 760.906652] env[61243]: INFO nova.compute.manager [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Terminating instance [ 760.907538] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Acquiring lock "refresh_cache-0d0504ae-8a82-4b5d-9d5d-535f30ceafa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.907705] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Acquired lock "refresh_cache-0d0504ae-8a82-4b5d-9d5d-535f30ceafa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.907865] env[61243]: DEBUG nova.network.neutron [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 761.194783] env[61243]: DEBUG oslo_concurrency.lockutils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.362s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.195355] env[61243]: DEBUG nova.compute.manager [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 761.198241] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.700s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.426424] env[61243]: DEBUG nova.network.neutron [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.511119] env[61243]: DEBUG nova.compute.manager [req-5ac9bbb5-9818-4cae-9fbf-deb62bf4a40e req-7bb62a19-abae-4259-b228-690572e27b04 service nova] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Received event network-changed-708bace8-628b-46fb-a8c8-4e9364b1824e {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.511119] env[61243]: DEBUG nova.compute.manager [req-5ac9bbb5-9818-4cae-9fbf-deb62bf4a40e req-7bb62a19-abae-4259-b228-690572e27b04 service nova] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Refreshing instance network info cache due to event network-changed-708bace8-628b-46fb-a8c8-4e9364b1824e. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 761.511119] env[61243]: DEBUG oslo_concurrency.lockutils [req-5ac9bbb5-9818-4cae-9fbf-deb62bf4a40e req-7bb62a19-abae-4259-b228-690572e27b04 service nova] Acquiring lock "refresh_cache-0d0504ae-8a82-4b5d-9d5d-535f30ceafa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.511595] env[61243]: DEBUG nova.network.neutron [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.702794] env[61243]: DEBUG nova.compute.utils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 761.707672] env[61243]: DEBUG nova.compute.manager [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 761.707865] env[61243]: DEBUG nova.network.neutron [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 761.748282] env[61243]: DEBUG nova.policy [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5ed217aea4542abb9f6fe77ed7e69f7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2e27639c13f424bbb03de483195a915', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 761.966772] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa7af25-62d3-4930-959e-9f3835f3c71f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.974555] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-661570e0-274d-4a3b-8387-f2cabe67da58 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.009862] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3faeb7-9f8e-4da1-af42-15cfea04f166 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.013125] env[61243]: DEBUG nova.network.neutron [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Successfully created port: 1912d33f-4a5c-43f1-9f74-7d3afaf17a72 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 762.016026] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Releasing lock "refresh_cache-0d0504ae-8a82-4b5d-9d5d-535f30ceafa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.016026] env[61243]: DEBUG nova.compute.manager [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 762.016026] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 762.016375] env[61243]: DEBUG oslo_concurrency.lockutils [req-5ac9bbb5-9818-4cae-9fbf-deb62bf4a40e req-7bb62a19-abae-4259-b228-690572e27b04 service nova] Acquired lock "refresh_cache-0d0504ae-8a82-4b5d-9d5d-535f30ceafa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.016580] env[61243]: DEBUG nova.network.neutron [req-5ac9bbb5-9818-4cae-9fbf-deb62bf4a40e req-7bb62a19-abae-4259-b228-690572e27b04 service nova] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Refreshing network info cache for port 708bace8-628b-46fb-a8c8-4e9364b1824e {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 762.017575] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-31cb16e4-0f80-4502-9e11-b4f416f36b1c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.023689] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c68ea516-9cd0-4bcc-a99c-3e1bd3c49085 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.031974] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7435a5-5ed5-4783-abcb-4ab416f7f8a1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.049197] env[61243]: DEBUG nova.compute.provider_tree [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.061583] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9 could not be found. [ 762.061796] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 762.061973] env[61243]: INFO nova.compute.manager [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Took 0.05 seconds to destroy the instance on the hypervisor. [ 762.062232] env[61243]: DEBUG oslo.service.loopingcall [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.062439] env[61243]: DEBUG nova.compute.manager [-] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.062537] env[61243]: DEBUG nova.network.neutron [-] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 762.081201] env[61243]: DEBUG nova.network.neutron [-] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.208343] env[61243]: DEBUG nova.compute.manager [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 762.542928] env[61243]: DEBUG nova.network.neutron [req-5ac9bbb5-9818-4cae-9fbf-deb62bf4a40e req-7bb62a19-abae-4259-b228-690572e27b04 service nova] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.552732] env[61243]: DEBUG nova.scheduler.client.report [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 762.585400] env[61243]: DEBUG nova.network.neutron [-] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.696071] env[61243]: DEBUG nova.network.neutron [req-5ac9bbb5-9818-4cae-9fbf-deb62bf4a40e req-7bb62a19-abae-4259-b228-690572e27b04 service nova] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.936969] env[61243]: ERROR nova.compute.manager [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1912d33f-4a5c-43f1-9f74-7d3afaf17a72, please check neutron logs for more information. [ 762.936969] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 762.936969] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.936969] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 762.936969] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.936969] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 762.936969] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.936969] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 762.936969] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.936969] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 762.936969] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.936969] env[61243]: ERROR nova.compute.manager raise self.value [ 762.936969] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.936969] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 762.936969] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.936969] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 762.937630] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.937630] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 762.937630] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1912d33f-4a5c-43f1-9f74-7d3afaf17a72, please check neutron logs for more information. [ 762.937630] env[61243]: ERROR nova.compute.manager [ 762.937630] env[61243]: Traceback (most recent call last): [ 762.937630] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 762.937630] env[61243]: listener.cb(fileno) [ 762.937630] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.937630] env[61243]: result = function(*args, **kwargs) [ 762.937630] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 762.937630] env[61243]: return func(*args, **kwargs) [ 762.937630] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.937630] env[61243]: raise e [ 762.937630] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.937630] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 762.937630] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.937630] env[61243]: created_port_ids = self._update_ports_for_instance( [ 762.937630] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.937630] env[61243]: with excutils.save_and_reraise_exception(): [ 762.937630] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.937630] env[61243]: self.force_reraise() [ 762.937630] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.937630] env[61243]: raise self.value [ 762.937630] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.937630] env[61243]: updated_port = self._update_port( [ 762.937630] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.937630] env[61243]: _ensure_no_port_binding_failure(port) [ 762.937630] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.937630] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 762.938547] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 1912d33f-4a5c-43f1-9f74-7d3afaf17a72, please check neutron logs for more information. [ 762.938547] env[61243]: Removing descriptor: 15 [ 763.058439] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.860s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.058840] env[61243]: ERROR nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 54ba2a00-34fa-4c51-941a-8b4175edb70c, please check neutron logs for more information. [ 763.058840] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] Traceback (most recent call last): [ 763.058840] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 763.058840] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] self.driver.spawn(context, instance, image_meta, [ 763.058840] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 763.058840] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 763.058840] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 763.058840] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] vm_ref = self.build_virtual_machine(instance, [ 763.058840] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 763.058840] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] vif_infos = vmwarevif.get_vif_info(self._session, [ 763.058840] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 763.059933] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] for vif in network_info: [ 763.059933] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 763.059933] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] return self._sync_wrapper(fn, *args, **kwargs) [ 763.059933] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 763.059933] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] self.wait() [ 763.059933] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 763.059933] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] self[:] = self._gt.wait() [ 763.059933] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 763.059933] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] return self._exit_event.wait() [ 763.059933] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 763.059933] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] current.throw(*self._exc) [ 763.059933] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 763.059933] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] result = function(*args, **kwargs) [ 763.060512] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 763.060512] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] return func(*args, **kwargs) [ 763.060512] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 763.060512] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] raise e [ 763.060512] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 763.060512] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] nwinfo = self.network_api.allocate_for_instance( [ 763.060512] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 763.060512] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] created_port_ids = self._update_ports_for_instance( [ 763.060512] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 763.060512] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] with excutils.save_and_reraise_exception(): [ 763.060512] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 763.060512] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] self.force_reraise() [ 763.060512] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 763.060954] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] raise self.value [ 763.060954] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 763.060954] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] updated_port = self._update_port( [ 763.060954] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 763.060954] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] _ensure_no_port_binding_failure(port) [ 763.060954] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 763.060954] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] raise exception.PortBindingFailed(port_id=port['id']) [ 763.060954] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] nova.exception.PortBindingFailed: Binding failed for port 54ba2a00-34fa-4c51-941a-8b4175edb70c, please check neutron logs for more information. [ 763.060954] env[61243]: ERROR nova.compute.manager [instance: e02f41be-125a-4228-8136-9e6639a268a9] [ 763.060954] env[61243]: DEBUG nova.compute.utils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Binding failed for port 54ba2a00-34fa-4c51-941a-8b4175edb70c, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 763.062183] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.484s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.064997] env[61243]: DEBUG nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Build of instance e02f41be-125a-4228-8136-9e6639a268a9 was re-scheduled: Binding failed for port 54ba2a00-34fa-4c51-941a-8b4175edb70c, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 763.065460] env[61243]: DEBUG nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 763.065691] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Acquiring lock "refresh_cache-e02f41be-125a-4228-8136-9e6639a268a9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.065839] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Acquired lock "refresh_cache-e02f41be-125a-4228-8136-9e6639a268a9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.065998] env[61243]: DEBUG nova.network.neutron [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 763.087630] env[61243]: INFO nova.compute.manager [-] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Took 1.02 seconds to deallocate network for instance. [ 763.089677] env[61243]: DEBUG nova.compute.claims [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 763.089859] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.199060] env[61243]: DEBUG oslo_concurrency.lockutils [req-5ac9bbb5-9818-4cae-9fbf-deb62bf4a40e req-7bb62a19-abae-4259-b228-690572e27b04 service nova] Releasing lock "refresh_cache-0d0504ae-8a82-4b5d-9d5d-535f30ceafa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.199060] env[61243]: DEBUG nova.compute.manager [req-5ac9bbb5-9818-4cae-9fbf-deb62bf4a40e req-7bb62a19-abae-4259-b228-690572e27b04 service nova] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Received event network-vif-deleted-708bace8-628b-46fb-a8c8-4e9364b1824e {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 763.217451] env[61243]: DEBUG nova.compute.manager [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 763.248485] env[61243]: DEBUG nova.virt.hardware [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 763.248726] env[61243]: DEBUG nova.virt.hardware [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 763.248895] env[61243]: DEBUG nova.virt.hardware [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 763.249107] env[61243]: DEBUG nova.virt.hardware [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 763.249294] env[61243]: DEBUG nova.virt.hardware [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 763.249457] env[61243]: DEBUG nova.virt.hardware [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 763.249662] env[61243]: DEBUG nova.virt.hardware [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 763.249820] env[61243]: DEBUG nova.virt.hardware [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 763.249984] env[61243]: DEBUG nova.virt.hardware [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 763.250160] env[61243]: DEBUG nova.virt.hardware [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 763.250385] env[61243]: DEBUG nova.virt.hardware [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 763.251271] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87c1be7-959e-411d-8567-79cf432db307 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.259950] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffe6719a-d81b-4ee2-9f06-dc389ea70e4f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.275611] env[61243]: ERROR nova.compute.manager [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1912d33f-4a5c-43f1-9f74-7d3afaf17a72, please check neutron logs for more information. [ 763.275611] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Traceback (most recent call last): [ 763.275611] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 763.275611] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] yield resources [ 763.275611] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 763.275611] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] self.driver.spawn(context, instance, image_meta, [ 763.275611] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 763.275611] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 763.275611] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 763.275611] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] vm_ref = self.build_virtual_machine(instance, [ 763.275611] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 763.276016] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] vif_infos = vmwarevif.get_vif_info(self._session, [ 763.276016] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 763.276016] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] for vif in network_info: [ 763.276016] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 763.276016] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] return self._sync_wrapper(fn, *args, **kwargs) [ 763.276016] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 763.276016] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] self.wait() [ 763.276016] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 763.276016] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] self[:] = self._gt.wait() [ 763.276016] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 763.276016] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] return self._exit_event.wait() [ 763.276016] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 763.276016] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] current.throw(*self._exc) [ 763.276535] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 763.276535] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] result = function(*args, **kwargs) [ 763.276535] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 763.276535] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] return func(*args, **kwargs) [ 763.276535] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 763.276535] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] raise e [ 763.276535] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 763.276535] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] nwinfo = self.network_api.allocate_for_instance( [ 763.276535] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 763.276535] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] created_port_ids = self._update_ports_for_instance( [ 763.276535] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 763.276535] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] with excutils.save_and_reraise_exception(): [ 763.276535] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 763.277133] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] self.force_reraise() [ 763.277133] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 763.277133] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] raise self.value [ 763.277133] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 763.277133] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] updated_port = self._update_port( [ 763.277133] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 763.277133] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] _ensure_no_port_binding_failure(port) [ 763.277133] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 763.277133] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] raise exception.PortBindingFailed(port_id=port['id']) [ 763.277133] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] nova.exception.PortBindingFailed: Binding failed for port 1912d33f-4a5c-43f1-9f74-7d3afaf17a72, please check neutron logs for more information. [ 763.277133] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] [ 763.277133] env[61243]: INFO nova.compute.manager [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Terminating instance [ 763.278084] env[61243]: DEBUG oslo_concurrency.lockutils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Acquiring lock "refresh_cache-2a81aa94-3e30-4f0d-b540-1f98cfa4cea9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.278261] env[61243]: DEBUG oslo_concurrency.lockutils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Acquired lock "refresh_cache-2a81aa94-3e30-4f0d-b540-1f98cfa4cea9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.278430] env[61243]: DEBUG nova.network.neutron [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 763.535809] env[61243]: DEBUG nova.compute.manager [req-8e82011a-6e1b-4d49-a1fa-19d047234b81 req-425f1008-de7e-47a8-a7bf-f05d1382069e service nova] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Received event network-changed-1912d33f-4a5c-43f1-9f74-7d3afaf17a72 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 763.535995] env[61243]: DEBUG nova.compute.manager [req-8e82011a-6e1b-4d49-a1fa-19d047234b81 req-425f1008-de7e-47a8-a7bf-f05d1382069e service nova] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Refreshing instance network info cache due to event network-changed-1912d33f-4a5c-43f1-9f74-7d3afaf17a72. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 763.536204] env[61243]: DEBUG oslo_concurrency.lockutils [req-8e82011a-6e1b-4d49-a1fa-19d047234b81 req-425f1008-de7e-47a8-a7bf-f05d1382069e service nova] Acquiring lock "refresh_cache-2a81aa94-3e30-4f0d-b540-1f98cfa4cea9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.586352] env[61243]: DEBUG nova.network.neutron [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.673392] env[61243]: DEBUG nova.network.neutron [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.795209] env[61243]: DEBUG nova.network.neutron [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.821688] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efaa6b18-2f7c-4a27-a08b-5d143b9b0489 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.829635] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a17fa325-0a6d-49d9-bffd-8b71a29f76f0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.860360] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c0541fb-09f6-49c7-a971-1c1b563ae44d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.868013] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2726757-869c-4a13-a8fc-bd5ec66880d3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.882609] env[61243]: DEBUG nova.compute.provider_tree [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.897386] env[61243]: DEBUG nova.network.neutron [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.176263] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Releasing lock "refresh_cache-e02f41be-125a-4228-8136-9e6639a268a9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.176555] env[61243]: DEBUG nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 764.176732] env[61243]: DEBUG nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 764.176957] env[61243]: DEBUG nova.network.neutron [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 764.192333] env[61243]: DEBUG nova.network.neutron [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.386151] env[61243]: DEBUG nova.scheduler.client.report [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.399577] env[61243]: DEBUG oslo_concurrency.lockutils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Releasing lock "refresh_cache-2a81aa94-3e30-4f0d-b540-1f98cfa4cea9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.399970] env[61243]: DEBUG nova.compute.manager [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 764.400173] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 764.400478] env[61243]: DEBUG oslo_concurrency.lockutils [req-8e82011a-6e1b-4d49-a1fa-19d047234b81 req-425f1008-de7e-47a8-a7bf-f05d1382069e service nova] Acquired lock "refresh_cache-2a81aa94-3e30-4f0d-b540-1f98cfa4cea9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.400647] env[61243]: DEBUG nova.network.neutron [req-8e82011a-6e1b-4d49-a1fa-19d047234b81 req-425f1008-de7e-47a8-a7bf-f05d1382069e service nova] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Refreshing network info cache for port 1912d33f-4a5c-43f1-9f74-7d3afaf17a72 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 764.404304] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-feeca7c9-651f-4d61-bf23-4cf7bf48f2a2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.411427] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af98910-e7ce-4aab-93e4-4a3607e150b4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.435267] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9 could not be found. [ 764.435507] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 764.435685] env[61243]: INFO nova.compute.manager [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 764.435919] env[61243]: DEBUG oslo.service.loopingcall [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.436164] env[61243]: DEBUG nova.compute.manager [-] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 764.436272] env[61243]: DEBUG nova.network.neutron [-] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 764.451787] env[61243]: DEBUG nova.network.neutron [-] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.695853] env[61243]: DEBUG nova.network.neutron [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.892052] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.830s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.892562] env[61243]: ERROR nova.compute.manager [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1a7dac77-793c-46dc-a760-ec1986c26daf, please check neutron logs for more information. [ 764.892562] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Traceback (most recent call last): [ 764.892562] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 764.892562] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] self.driver.spawn(context, instance, image_meta, [ 764.892562] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 764.892562] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] self._vmops.spawn(context, instance, image_meta, injected_files, [ 764.892562] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 764.892562] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] vm_ref = self.build_virtual_machine(instance, [ 764.892562] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 764.892562] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] vif_infos = vmwarevif.get_vif_info(self._session, [ 764.892562] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 764.893052] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] for vif in network_info: [ 764.893052] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 764.893052] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] return self._sync_wrapper(fn, *args, **kwargs) [ 764.893052] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 764.893052] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] self.wait() [ 764.893052] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 764.893052] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] self[:] = self._gt.wait() [ 764.893052] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 764.893052] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] return self._exit_event.wait() [ 764.893052] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 764.893052] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] current.throw(*self._exc) [ 764.893052] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.893052] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] result = function(*args, **kwargs) [ 764.893511] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 764.893511] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] return func(*args, **kwargs) [ 764.893511] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.893511] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] raise e [ 764.893511] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.893511] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] nwinfo = self.network_api.allocate_for_instance( [ 764.893511] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.893511] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] created_port_ids = self._update_ports_for_instance( [ 764.893511] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.893511] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] with excutils.save_and_reraise_exception(): [ 764.893511] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.893511] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] self.force_reraise() [ 764.893511] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.893930] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] raise self.value [ 764.893930] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.893930] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] updated_port = self._update_port( [ 764.893930] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.893930] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] _ensure_no_port_binding_failure(port) [ 764.893930] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.893930] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] raise exception.PortBindingFailed(port_id=port['id']) [ 764.893930] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] nova.exception.PortBindingFailed: Binding failed for port 1a7dac77-793c-46dc-a760-ec1986c26daf, please check neutron logs for more information. [ 764.893930] env[61243]: ERROR nova.compute.manager [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] [ 764.893930] env[61243]: DEBUG nova.compute.utils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Binding failed for port 1a7dac77-793c-46dc-a760-ec1986c26daf, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 764.894718] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.989s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.896515] env[61243]: INFO nova.compute.claims [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 764.899078] env[61243]: DEBUG nova.compute.manager [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Build of instance b1bad82d-d7bf-4e58-bbf2-f3d68f214115 was re-scheduled: Binding failed for port 1a7dac77-793c-46dc-a760-ec1986c26daf, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 764.899495] env[61243]: DEBUG nova.compute.manager [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 764.899728] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Acquiring lock "refresh_cache-b1bad82d-d7bf-4e58-bbf2-f3d68f214115" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.899876] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Acquired lock "refresh_cache-b1bad82d-d7bf-4e58-bbf2-f3d68f214115" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.900044] env[61243]: DEBUG nova.network.neutron [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 764.920064] env[61243]: DEBUG nova.network.neutron [req-8e82011a-6e1b-4d49-a1fa-19d047234b81 req-425f1008-de7e-47a8-a7bf-f05d1382069e service nova] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.955311] env[61243]: DEBUG nova.network.neutron [-] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.991658] env[61243]: DEBUG nova.network.neutron [req-8e82011a-6e1b-4d49-a1fa-19d047234b81 req-425f1008-de7e-47a8-a7bf-f05d1382069e service nova] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.198099] env[61243]: INFO nova.compute.manager [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] [instance: e02f41be-125a-4228-8136-9e6639a268a9] Took 1.02 seconds to deallocate network for instance. [ 765.420215] env[61243]: DEBUG nova.network.neutron [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.459664] env[61243]: INFO nova.compute.manager [-] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Took 1.02 seconds to deallocate network for instance. [ 765.461888] env[61243]: DEBUG nova.compute.claims [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 765.462098] env[61243]: DEBUG oslo_concurrency.lockutils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.463187] env[61243]: DEBUG nova.network.neutron [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.494145] env[61243]: DEBUG oslo_concurrency.lockutils [req-8e82011a-6e1b-4d49-a1fa-19d047234b81 req-425f1008-de7e-47a8-a7bf-f05d1382069e service nova] Releasing lock "refresh_cache-2a81aa94-3e30-4f0d-b540-1f98cfa4cea9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.494339] env[61243]: DEBUG nova.compute.manager [req-8e82011a-6e1b-4d49-a1fa-19d047234b81 req-425f1008-de7e-47a8-a7bf-f05d1382069e service nova] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Received event network-vif-deleted-1912d33f-4a5c-43f1-9f74-7d3afaf17a72 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 765.966135] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Releasing lock "refresh_cache-b1bad82d-d7bf-4e58-bbf2-f3d68f214115" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.966135] env[61243]: DEBUG nova.compute.manager [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 765.966319] env[61243]: DEBUG nova.compute.manager [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 765.966506] env[61243]: DEBUG nova.network.neutron [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 765.982148] env[61243]: DEBUG nova.network.neutron [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.129543] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d49330a4-c716-45ae-83cf-ff7580f1156b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.137398] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb2acaf-78e4-4210-af0a-79ff63061641 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.167311] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb6f97c-d0d8-4970-99fb-eddcfd0a43e5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.174922] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9841927-cb93-457f-969b-3ff9543a7d52 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.187870] env[61243]: DEBUG nova.compute.provider_tree [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.226359] env[61243]: INFO nova.scheduler.client.report [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Deleted allocations for instance e02f41be-125a-4228-8136-9e6639a268a9 [ 766.484184] env[61243]: DEBUG nova.network.neutron [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.690987] env[61243]: DEBUG nova.scheduler.client.report [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.733924] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0e6c3a01-b5eb-493d-96ea-e6a56f0dd972 tempest-ServersTestBootFromVolume-392190390 tempest-ServersTestBootFromVolume-392190390-project-member] Lock "e02f41be-125a-4228-8136-9e6639a268a9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 178.397s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.987545] env[61243]: INFO nova.compute.manager [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] [instance: b1bad82d-d7bf-4e58-bbf2-f3d68f214115] Took 1.02 seconds to deallocate network for instance. [ 767.197190] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.197190] env[61243]: DEBUG nova.compute.manager [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 767.199737] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.689s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.235849] env[61243]: DEBUG nova.compute.manager [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 767.710026] env[61243]: DEBUG nova.compute.utils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 767.711696] env[61243]: DEBUG nova.compute.manager [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 767.712089] env[61243]: DEBUG nova.network.neutron [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 767.760183] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.771137] env[61243]: DEBUG nova.policy [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1d64f20db125476a9e2cf5a77c88437e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f636baf6c50d49919d0c3286e83192b3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 767.984009] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d07e43-c8fa-4147-88db-1ec05993b487 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.991750] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed68efb-06ff-4481-b433-a46017a653f4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.025851] env[61243]: INFO nova.scheduler.client.report [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Deleted allocations for instance b1bad82d-d7bf-4e58-bbf2-f3d68f214115 [ 768.032517] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed1309ff-8ded-4c8c-90e9-47fcc430470a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.040836] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c73ad43-37ee-4def-b8b2-2bc62c0ec805 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.055027] env[61243]: DEBUG nova.compute.provider_tree [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.078385] env[61243]: DEBUG nova.network.neutron [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Successfully created port: 0bcb6106-50de-4d90-80ac-ce28ca91123a {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 768.215200] env[61243]: DEBUG nova.compute.manager [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 768.345096] env[61243]: DEBUG nova.network.neutron [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Successfully created port: 876f6efb-8bc0-4994-8555-bedb2ff4e242 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 768.540108] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d84a7414-3b19-4ad2-a7f3-49afecf3f4bf tempest-MigrationsAdminTest-1746581089 tempest-MigrationsAdminTest-1746581089-project-member] Lock "b1bad82d-d7bf-4e58-bbf2-f3d68f214115" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 179.021s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.557902] env[61243]: DEBUG nova.scheduler.client.report [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 769.043797] env[61243]: DEBUG nova.compute.manager [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.070046] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.869s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.070046] env[61243]: ERROR nova.compute.manager [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8f597029-2d2c-4f15-8b33-593740cd82d6, please check neutron logs for more information. [ 769.070046] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Traceback (most recent call last): [ 769.070046] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 769.070046] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] self.driver.spawn(context, instance, image_meta, [ 769.070046] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 769.070046] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.070046] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.070046] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] vm_ref = self.build_virtual_machine(instance, [ 769.070795] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.070795] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.070795] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.070795] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] for vif in network_info: [ 769.070795] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 769.070795] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] return self._sync_wrapper(fn, *args, **kwargs) [ 769.070795] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 769.070795] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] self.wait() [ 769.070795] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 769.070795] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] self[:] = self._gt.wait() [ 769.070795] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.070795] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] return self._exit_event.wait() [ 769.070795] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 769.071340] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] current.throw(*self._exc) [ 769.071340] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.071340] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] result = function(*args, **kwargs) [ 769.071340] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 769.071340] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] return func(*args, **kwargs) [ 769.071340] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.071340] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] raise e [ 769.071340] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.071340] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] nwinfo = self.network_api.allocate_for_instance( [ 769.071340] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.071340] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] created_port_ids = self._update_ports_for_instance( [ 769.071340] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.071340] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] with excutils.save_and_reraise_exception(): [ 769.071809] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.071809] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] self.force_reraise() [ 769.071809] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.071809] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] raise self.value [ 769.071809] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.071809] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] updated_port = self._update_port( [ 769.071809] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.071809] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] _ensure_no_port_binding_failure(port) [ 769.071809] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.071809] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] raise exception.PortBindingFailed(port_id=port['id']) [ 769.071809] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] nova.exception.PortBindingFailed: Binding failed for port 8f597029-2d2c-4f15-8b33-593740cd82d6, please check neutron logs for more information. [ 769.071809] env[61243]: ERROR nova.compute.manager [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] [ 769.072263] env[61243]: DEBUG nova.compute.utils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Binding failed for port 8f597029-2d2c-4f15-8b33-593740cd82d6, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 769.077252] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.049s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.077252] env[61243]: INFO nova.compute.claims [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 769.081845] env[61243]: DEBUG nova.compute.manager [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Build of instance 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20 was re-scheduled: Binding failed for port 8f597029-2d2c-4f15-8b33-593740cd82d6, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 769.082468] env[61243]: DEBUG nova.compute.manager [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 769.082567] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Acquiring lock "refresh_cache-61d68f1c-6a48-475a-8cf5-fa6ce3b09f20" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.082777] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Acquired lock "refresh_cache-61d68f1c-6a48-475a-8cf5-fa6ce3b09f20" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.082907] env[61243]: DEBUG nova.network.neutron [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 769.226805] env[61243]: DEBUG nova.compute.manager [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 769.268719] env[61243]: DEBUG nova.virt.hardware [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 769.269034] env[61243]: DEBUG nova.virt.hardware [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 769.269120] env[61243]: DEBUG nova.virt.hardware [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 769.269357] env[61243]: DEBUG nova.virt.hardware [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 769.269443] env[61243]: DEBUG nova.virt.hardware [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 769.270093] env[61243]: DEBUG nova.virt.hardware [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 769.270093] env[61243]: DEBUG nova.virt.hardware [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 769.270093] env[61243]: DEBUG nova.virt.hardware [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 769.273990] env[61243]: DEBUG nova.virt.hardware [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 769.274251] env[61243]: DEBUG nova.virt.hardware [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 769.274864] env[61243]: DEBUG nova.virt.hardware [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 769.275594] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c3a8ca-a2e4-40f3-8db2-f01559ca8901 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.283757] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e61e8fcf-daa6-443d-a653-b9caa5b44bef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.341838] env[61243]: DEBUG nova.compute.manager [req-da386e23-957f-4994-8d22-21786bcc4803 req-38039289-0d03-46d1-ad34-7032f08662e2 service nova] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Received event network-changed-0bcb6106-50de-4d90-80ac-ce28ca91123a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.342025] env[61243]: DEBUG nova.compute.manager [req-da386e23-957f-4994-8d22-21786bcc4803 req-38039289-0d03-46d1-ad34-7032f08662e2 service nova] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Refreshing instance network info cache due to event network-changed-0bcb6106-50de-4d90-80ac-ce28ca91123a. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 769.342246] env[61243]: DEBUG oslo_concurrency.lockutils [req-da386e23-957f-4994-8d22-21786bcc4803 req-38039289-0d03-46d1-ad34-7032f08662e2 service nova] Acquiring lock "refresh_cache-1f4985a7-f6cc-4e79-aea3-ce0623aed515" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.342426] env[61243]: DEBUG oslo_concurrency.lockutils [req-da386e23-957f-4994-8d22-21786bcc4803 req-38039289-0d03-46d1-ad34-7032f08662e2 service nova] Acquired lock "refresh_cache-1f4985a7-f6cc-4e79-aea3-ce0623aed515" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.342623] env[61243]: DEBUG nova.network.neutron [req-da386e23-957f-4994-8d22-21786bcc4803 req-38039289-0d03-46d1-ad34-7032f08662e2 service nova] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Refreshing network info cache for port 0bcb6106-50de-4d90-80ac-ce28ca91123a {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 769.570154] env[61243]: DEBUG oslo_concurrency.lockutils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.614546] env[61243]: DEBUG nova.network.neutron [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.738332] env[61243]: DEBUG nova.network.neutron [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.769778] env[61243]: ERROR nova.compute.manager [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0bcb6106-50de-4d90-80ac-ce28ca91123a, please check neutron logs for more information. [ 769.769778] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 769.769778] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.769778] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 769.769778] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.769778] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 769.769778] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.769778] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 769.769778] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.769778] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 769.769778] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.769778] env[61243]: ERROR nova.compute.manager raise self.value [ 769.769778] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.769778] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 769.769778] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.769778] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 769.770316] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.770316] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 769.770316] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0bcb6106-50de-4d90-80ac-ce28ca91123a, please check neutron logs for more information. [ 769.770316] env[61243]: ERROR nova.compute.manager [ 769.770316] env[61243]: Traceback (most recent call last): [ 769.770316] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 769.770316] env[61243]: listener.cb(fileno) [ 769.770316] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.770316] env[61243]: result = function(*args, **kwargs) [ 769.770316] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 769.770316] env[61243]: return func(*args, **kwargs) [ 769.770316] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.770316] env[61243]: raise e [ 769.770316] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.770316] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 769.770316] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.770316] env[61243]: created_port_ids = self._update_ports_for_instance( [ 769.770316] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.770316] env[61243]: with excutils.save_and_reraise_exception(): [ 769.770316] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.770316] env[61243]: self.force_reraise() [ 769.770316] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.770316] env[61243]: raise self.value [ 769.770316] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.770316] env[61243]: updated_port = self._update_port( [ 769.770316] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.770316] env[61243]: _ensure_no_port_binding_failure(port) [ 769.770316] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.770316] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 769.771188] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 0bcb6106-50de-4d90-80ac-ce28ca91123a, please check neutron logs for more information. [ 769.771188] env[61243]: Removing descriptor: 15 [ 769.771188] env[61243]: ERROR nova.compute.manager [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0bcb6106-50de-4d90-80ac-ce28ca91123a, please check neutron logs for more information. [ 769.771188] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Traceback (most recent call last): [ 769.771188] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 769.771188] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] yield resources [ 769.771188] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 769.771188] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] self.driver.spawn(context, instance, image_meta, [ 769.771188] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 769.771188] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.771188] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.771188] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] vm_ref = self.build_virtual_machine(instance, [ 769.771656] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.771656] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.771656] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.771656] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] for vif in network_info: [ 769.771656] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 769.771656] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] return self._sync_wrapper(fn, *args, **kwargs) [ 769.771656] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 769.771656] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] self.wait() [ 769.771656] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 769.771656] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] self[:] = self._gt.wait() [ 769.771656] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.771656] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] return self._exit_event.wait() [ 769.771656] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 769.772084] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] result = hub.switch() [ 769.772084] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 769.772084] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] return self.greenlet.switch() [ 769.772084] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.772084] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] result = function(*args, **kwargs) [ 769.772084] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 769.772084] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] return func(*args, **kwargs) [ 769.772084] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.772084] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] raise e [ 769.772084] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.772084] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] nwinfo = self.network_api.allocate_for_instance( [ 769.772084] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.772084] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] created_port_ids = self._update_ports_for_instance( [ 769.772497] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.772497] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] with excutils.save_and_reraise_exception(): [ 769.772497] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.772497] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] self.force_reraise() [ 769.772497] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.772497] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] raise self.value [ 769.772497] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.772497] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] updated_port = self._update_port( [ 769.772497] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.772497] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] _ensure_no_port_binding_failure(port) [ 769.772497] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.772497] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] raise exception.PortBindingFailed(port_id=port['id']) [ 769.772869] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] nova.exception.PortBindingFailed: Binding failed for port 0bcb6106-50de-4d90-80ac-ce28ca91123a, please check neutron logs for more information. [ 769.772869] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] [ 769.772869] env[61243]: INFO nova.compute.manager [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Terminating instance [ 769.774670] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Acquiring lock "refresh_cache-1f4985a7-f6cc-4e79-aea3-ce0623aed515" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.874792] env[61243]: DEBUG nova.network.neutron [req-da386e23-957f-4994-8d22-21786bcc4803 req-38039289-0d03-46d1-ad34-7032f08662e2 service nova] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.961273] env[61243]: DEBUG nova.network.neutron [req-da386e23-957f-4994-8d22-21786bcc4803 req-38039289-0d03-46d1-ad34-7032f08662e2 service nova] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.244619] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Releasing lock "refresh_cache-61d68f1c-6a48-475a-8cf5-fa6ce3b09f20" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.244780] env[61243]: DEBUG nova.compute.manager [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 770.244967] env[61243]: DEBUG nova.compute.manager [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 770.245155] env[61243]: DEBUG nova.network.neutron [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 770.264789] env[61243]: DEBUG nova.network.neutron [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.304085] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bfe40fe-ca88-4c05-ac8d-43c1cde80074 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.313240] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca13b49-2b6a-4aaf-8445-988d93ec370d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.343294] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c75b95aa-8149-4ddf-960b-b45425ddcf96 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.350984] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5c7acb-a607-4280-9751-ca8dc40262c0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.364197] env[61243]: DEBUG nova.compute.provider_tree [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.464416] env[61243]: DEBUG oslo_concurrency.lockutils [req-da386e23-957f-4994-8d22-21786bcc4803 req-38039289-0d03-46d1-ad34-7032f08662e2 service nova] Releasing lock "refresh_cache-1f4985a7-f6cc-4e79-aea3-ce0623aed515" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.464589] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Acquired lock "refresh_cache-1f4985a7-f6cc-4e79-aea3-ce0623aed515" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.465425] env[61243]: DEBUG nova.network.neutron [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 770.767208] env[61243]: DEBUG nova.network.neutron [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.867301] env[61243]: DEBUG nova.scheduler.client.report [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.014576] env[61243]: DEBUG nova.network.neutron [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.106059] env[61243]: DEBUG nova.network.neutron [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.271568] env[61243]: INFO nova.compute.manager [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] [instance: 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20] Took 1.03 seconds to deallocate network for instance. [ 771.282525] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "729d9bd9-1ab0-47bd-9b7a-11412d3dc608" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.282740] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "729d9bd9-1ab0-47bd-9b7a-11412d3dc608" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.368152] env[61243]: DEBUG nova.compute.manager [req-3a6a0665-7c49-43f3-89c3-a7e91d2a5088 req-bd150bc8-70b8-4ce1-b946-c87898c72bd2 service nova] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Received event network-vif-deleted-0bcb6106-50de-4d90-80ac-ce28ca91123a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.375364] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.303s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.375847] env[61243]: DEBUG nova.compute.manager [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 771.378220] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 14.780s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.378375] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.378535] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61243) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 771.378803] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.696s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.382150] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15840de9-d5ed-4489-bb24-3c7eea258bb0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.390899] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61a32c83-d1f9-4309-b40a-3feb33ee070e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.406685] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2acc1c-7d93-4171-9e56-be2a01e0bdfc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.413630] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2c4501-fcfa-4bf0-aba0-6fc0394ff841 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.445915] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181516MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=61243) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 771.446093] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.493109] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "96315d9e-4eda-4e3a-af0d-bdc52ab181e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.493375] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "96315d9e-4eda-4e3a-af0d-bdc52ab181e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.608151] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Releasing lock "refresh_cache-1f4985a7-f6cc-4e79-aea3-ce0623aed515" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.608620] env[61243]: DEBUG nova.compute.manager [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 771.608765] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 771.609073] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2cb3f9a3-1050-43e1-a646-c5bfd3480d75 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.618633] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c1979f1-4b01-427c-8145-10d08ac8e971 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.641020] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1f4985a7-f6cc-4e79-aea3-ce0623aed515 could not be found. [ 771.641020] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 771.641020] env[61243]: INFO nova.compute.manager [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Took 0.03 seconds to destroy the instance on the hypervisor. [ 771.641020] env[61243]: DEBUG oslo.service.loopingcall [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.641240] env[61243]: DEBUG nova.compute.manager [-] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.641240] env[61243]: DEBUG nova.network.neutron [-] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 771.696268] env[61243]: DEBUG nova.network.neutron [-] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.882984] env[61243]: DEBUG nova.compute.utils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 771.884439] env[61243]: DEBUG nova.compute.manager [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 771.884614] env[61243]: DEBUG nova.network.neutron [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 771.943728] env[61243]: DEBUG nova.policy [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a5f01f350674f688c3ccad04b6bea03', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e5591d7a03144f7ea1f566f3785aa06a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 772.191969] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d753053-db23-480f-b3c5-16b8dd6627d7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.200234] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86db7e6d-2828-4450-8516-fbd2f44de448 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.230887] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66104c18-d3be-4757-bdd5-1a7860eb59c8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.239055] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7062da53-6000-4ebb-b561-179fa0d56031 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.255311] env[61243]: DEBUG nova.compute.provider_tree [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.260802] env[61243]: DEBUG nova.network.neutron [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Successfully created port: 2c60b335-f40e-41b2-9f29-07d59e2562c6 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 772.305792] env[61243]: INFO nova.scheduler.client.report [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Deleted allocations for instance 61d68f1c-6a48-475a-8cf5-fa6ce3b09f20 [ 772.387503] env[61243]: DEBUG nova.compute.manager [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 772.425196] env[61243]: DEBUG nova.network.neutron [-] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.767042] env[61243]: DEBUG nova.scheduler.client.report [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 772.814553] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d86a3341-03f1-430f-9161-6c6fcc5f6087 tempest-SecurityGroupsTestJSON-487491325 tempest-SecurityGroupsTestJSON-487491325-project-member] Lock "61d68f1c-6a48-475a-8cf5-fa6ce3b09f20" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 167.112s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.930061] env[61243]: INFO nova.compute.manager [-] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Took 1.29 seconds to deallocate network for instance. [ 772.932456] env[61243]: DEBUG nova.compute.claims [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 772.932595] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.201045] env[61243]: ERROR nova.compute.manager [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2c60b335-f40e-41b2-9f29-07d59e2562c6, please check neutron logs for more information. [ 773.201045] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 773.201045] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.201045] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 773.201045] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.201045] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 773.201045] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.201045] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 773.201045] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.201045] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 773.201045] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.201045] env[61243]: ERROR nova.compute.manager raise self.value [ 773.201045] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.201045] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 773.201045] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.201045] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 773.201647] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.201647] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 773.201647] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2c60b335-f40e-41b2-9f29-07d59e2562c6, please check neutron logs for more information. [ 773.201647] env[61243]: ERROR nova.compute.manager [ 773.201647] env[61243]: Traceback (most recent call last): [ 773.201647] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 773.201647] env[61243]: listener.cb(fileno) [ 773.201647] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.201647] env[61243]: result = function(*args, **kwargs) [ 773.201647] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.201647] env[61243]: return func(*args, **kwargs) [ 773.201647] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.201647] env[61243]: raise e [ 773.201647] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.201647] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 773.201647] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.201647] env[61243]: created_port_ids = self._update_ports_for_instance( [ 773.201647] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.201647] env[61243]: with excutils.save_and_reraise_exception(): [ 773.201647] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.201647] env[61243]: self.force_reraise() [ 773.201647] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.201647] env[61243]: raise self.value [ 773.201647] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.201647] env[61243]: updated_port = self._update_port( [ 773.201647] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.201647] env[61243]: _ensure_no_port_binding_failure(port) [ 773.201647] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.201647] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 773.202582] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 2c60b335-f40e-41b2-9f29-07d59e2562c6, please check neutron logs for more information. [ 773.202582] env[61243]: Removing descriptor: 16 [ 773.269441] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.890s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.270063] env[61243]: ERROR nova.compute.manager [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2363167a-7627-41e4-9d85-beb565111760, please check neutron logs for more information. [ 773.270063] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Traceback (most recent call last): [ 773.270063] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 773.270063] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] self.driver.spawn(context, instance, image_meta, [ 773.270063] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 773.270063] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 773.270063] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 773.270063] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] vm_ref = self.build_virtual_machine(instance, [ 773.270063] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 773.270063] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] vif_infos = vmwarevif.get_vif_info(self._session, [ 773.270063] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 773.270431] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] for vif in network_info: [ 773.270431] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 773.270431] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] return self._sync_wrapper(fn, *args, **kwargs) [ 773.270431] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 773.270431] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] self.wait() [ 773.270431] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 773.270431] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] self[:] = self._gt.wait() [ 773.270431] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 773.270431] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] return self._exit_event.wait() [ 773.270431] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 773.270431] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] current.throw(*self._exc) [ 773.270431] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.270431] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] result = function(*args, **kwargs) [ 773.270958] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.270958] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] return func(*args, **kwargs) [ 773.270958] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.270958] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] raise e [ 773.270958] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.270958] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] nwinfo = self.network_api.allocate_for_instance( [ 773.270958] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.270958] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] created_port_ids = self._update_ports_for_instance( [ 773.270958] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.270958] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] with excutils.save_and_reraise_exception(): [ 773.270958] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.270958] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] self.force_reraise() [ 773.270958] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.271359] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] raise self.value [ 773.271359] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.271359] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] updated_port = self._update_port( [ 773.271359] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.271359] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] _ensure_no_port_binding_failure(port) [ 773.271359] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.271359] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] raise exception.PortBindingFailed(port_id=port['id']) [ 773.271359] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] nova.exception.PortBindingFailed: Binding failed for port 2363167a-7627-41e4-9d85-beb565111760, please check neutron logs for more information. [ 773.271359] env[61243]: ERROR nova.compute.manager [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] [ 773.271359] env[61243]: DEBUG nova.compute.utils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Binding failed for port 2363167a-7627-41e4-9d85-beb565111760, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 773.272102] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.875s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.275405] env[61243]: DEBUG nova.compute.manager [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Build of instance 454d97a2-40d0-4c7a-9df7-04c391715e0d was re-scheduled: Binding failed for port 2363167a-7627-41e4-9d85-beb565111760, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 773.275838] env[61243]: DEBUG nova.compute.manager [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 773.276082] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Acquiring lock "refresh_cache-454d97a2-40d0-4c7a-9df7-04c391715e0d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.276235] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Acquired lock "refresh_cache-454d97a2-40d0-4c7a-9df7-04c391715e0d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.276396] env[61243]: DEBUG nova.network.neutron [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 773.317704] env[61243]: DEBUG nova.compute.manager [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 773.397864] env[61243]: DEBUG nova.compute.manager [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 773.424896] env[61243]: DEBUG nova.virt.hardware [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 773.425877] env[61243]: DEBUG nova.virt.hardware [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 773.426072] env[61243]: DEBUG nova.virt.hardware [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 773.426273] env[61243]: DEBUG nova.virt.hardware [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 773.426563] env[61243]: DEBUG nova.virt.hardware [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 773.426666] env[61243]: DEBUG nova.virt.hardware [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 773.426914] env[61243]: DEBUG nova.virt.hardware [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 773.427091] env[61243]: DEBUG nova.virt.hardware [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 773.427563] env[61243]: DEBUG nova.virt.hardware [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 773.427757] env[61243]: DEBUG nova.virt.hardware [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 773.427926] env[61243]: DEBUG nova.virt.hardware [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 773.428900] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdbe23eb-0959-4186-aa64-9002fb3f4bc8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.437634] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6dc59ef-7fc4-469b-8ef4-93c9d8f02b86 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.454554] env[61243]: ERROR nova.compute.manager [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2c60b335-f40e-41b2-9f29-07d59e2562c6, please check neutron logs for more information. [ 773.454554] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Traceback (most recent call last): [ 773.454554] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 773.454554] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] yield resources [ 773.454554] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 773.454554] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] self.driver.spawn(context, instance, image_meta, [ 773.454554] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 773.454554] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 773.454554] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 773.454554] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] vm_ref = self.build_virtual_machine(instance, [ 773.454554] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 773.454973] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] vif_infos = vmwarevif.get_vif_info(self._session, [ 773.454973] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 773.454973] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] for vif in network_info: [ 773.454973] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 773.454973] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] return self._sync_wrapper(fn, *args, **kwargs) [ 773.454973] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 773.454973] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] self.wait() [ 773.454973] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 773.454973] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] self[:] = self._gt.wait() [ 773.454973] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 773.454973] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] return self._exit_event.wait() [ 773.454973] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 773.454973] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] current.throw(*self._exc) [ 773.455435] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.455435] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] result = function(*args, **kwargs) [ 773.455435] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.455435] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] return func(*args, **kwargs) [ 773.455435] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.455435] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] raise e [ 773.455435] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.455435] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] nwinfo = self.network_api.allocate_for_instance( [ 773.455435] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.455435] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] created_port_ids = self._update_ports_for_instance( [ 773.455435] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.455435] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] with excutils.save_and_reraise_exception(): [ 773.455435] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.456021] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] self.force_reraise() [ 773.456021] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.456021] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] raise self.value [ 773.456021] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.456021] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] updated_port = self._update_port( [ 773.456021] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.456021] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] _ensure_no_port_binding_failure(port) [ 773.456021] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.456021] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] raise exception.PortBindingFailed(port_id=port['id']) [ 773.456021] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] nova.exception.PortBindingFailed: Binding failed for port 2c60b335-f40e-41b2-9f29-07d59e2562c6, please check neutron logs for more information. [ 773.456021] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] [ 773.456021] env[61243]: INFO nova.compute.manager [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Terminating instance [ 773.458160] env[61243]: DEBUG nova.compute.manager [req-2dc3295d-a9d1-4f89-a581-ec04bd369c36 req-17fbb186-a08f-4eea-ac09-d71ee2441cfe service nova] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Received event network-changed-2c60b335-f40e-41b2-9f29-07d59e2562c6 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.458347] env[61243]: DEBUG nova.compute.manager [req-2dc3295d-a9d1-4f89-a581-ec04bd369c36 req-17fbb186-a08f-4eea-ac09-d71ee2441cfe service nova] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Refreshing instance network info cache due to event network-changed-2c60b335-f40e-41b2-9f29-07d59e2562c6. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 773.458553] env[61243]: DEBUG oslo_concurrency.lockutils [req-2dc3295d-a9d1-4f89-a581-ec04bd369c36 req-17fbb186-a08f-4eea-ac09-d71ee2441cfe service nova] Acquiring lock "refresh_cache-5314f922-6603-40dd-8760-9c1b77d6bd0a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.458699] env[61243]: DEBUG oslo_concurrency.lockutils [req-2dc3295d-a9d1-4f89-a581-ec04bd369c36 req-17fbb186-a08f-4eea-ac09-d71ee2441cfe service nova] Acquired lock "refresh_cache-5314f922-6603-40dd-8760-9c1b77d6bd0a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.458851] env[61243]: DEBUG nova.network.neutron [req-2dc3295d-a9d1-4f89-a581-ec04bd369c36 req-17fbb186-a08f-4eea-ac09-d71ee2441cfe service nova] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Refreshing network info cache for port 2c60b335-f40e-41b2-9f29-07d59e2562c6 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 773.459947] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Acquiring lock "refresh_cache-5314f922-6603-40dd-8760-9c1b77d6bd0a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.697760] env[61243]: DEBUG oslo_concurrency.lockutils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Acquiring lock "6f33b303-1ecd-478e-8fa3-33a5f9a5af50" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.698027] env[61243]: DEBUG oslo_concurrency.lockutils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Lock "6f33b303-1ecd-478e-8fa3-33a5f9a5af50" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.808415] env[61243]: DEBUG nova.network.neutron [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.841415] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.897176] env[61243]: DEBUG nova.network.neutron [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.980577] env[61243]: DEBUG nova.network.neutron [req-2dc3295d-a9d1-4f89-a581-ec04bd369c36 req-17fbb186-a08f-4eea-ac09-d71ee2441cfe service nova] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.078395] env[61243]: DEBUG nova.network.neutron [req-2dc3295d-a9d1-4f89-a581-ec04bd369c36 req-17fbb186-a08f-4eea-ac09-d71ee2441cfe service nova] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.103320] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286d4791-7bf2-4298-bd31-36753f5954c1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.111153] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a525bf4f-b957-419e-a32a-a7fde5797541 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.143072] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c302a2-c1bf-4af1-b8cf-be3ce2940a23 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.149280] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d0e792c-2588-4d25-a01c-e047b33d2a44 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.163184] env[61243]: DEBUG nova.compute.provider_tree [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.400280] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Releasing lock "refresh_cache-454d97a2-40d0-4c7a-9df7-04c391715e0d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.400516] env[61243]: DEBUG nova.compute.manager [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 774.400699] env[61243]: DEBUG nova.compute.manager [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 774.400865] env[61243]: DEBUG nova.network.neutron [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 774.417418] env[61243]: DEBUG nova.network.neutron [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.583685] env[61243]: DEBUG oslo_concurrency.lockutils [req-2dc3295d-a9d1-4f89-a581-ec04bd369c36 req-17fbb186-a08f-4eea-ac09-d71ee2441cfe service nova] Releasing lock "refresh_cache-5314f922-6603-40dd-8760-9c1b77d6bd0a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.583956] env[61243]: DEBUG nova.compute.manager [req-2dc3295d-a9d1-4f89-a581-ec04bd369c36 req-17fbb186-a08f-4eea-ac09-d71ee2441cfe service nova] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Received event network-vif-deleted-2c60b335-f40e-41b2-9f29-07d59e2562c6 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 774.584335] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Acquired lock "refresh_cache-5314f922-6603-40dd-8760-9c1b77d6bd0a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.584506] env[61243]: DEBUG nova.network.neutron [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 774.666721] env[61243]: DEBUG nova.scheduler.client.report [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.921064] env[61243]: DEBUG nova.network.neutron [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.103262] env[61243]: DEBUG nova.network.neutron [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.172062] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.900s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.172683] env[61243]: ERROR nova.compute.manager [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f6efca39-d061-4332-a8c2-76d8cfc414b3, please check neutron logs for more information. [ 775.172683] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Traceback (most recent call last): [ 775.172683] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 775.172683] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] self.driver.spawn(context, instance, image_meta, [ 775.172683] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 775.172683] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 775.172683] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 775.172683] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] vm_ref = self.build_virtual_machine(instance, [ 775.172683] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 775.172683] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] vif_infos = vmwarevif.get_vif_info(self._session, [ 775.172683] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 775.172901] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] for vif in network_info: [ 775.172901] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 775.172901] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] return self._sync_wrapper(fn, *args, **kwargs) [ 775.172901] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 775.172901] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] self.wait() [ 775.172901] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 775.172901] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] self[:] = self._gt.wait() [ 775.172901] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 775.172901] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] return self._exit_event.wait() [ 775.172901] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 775.172901] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] result = hub.switch() [ 775.172901] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 775.172901] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] return self.greenlet.switch() [ 775.173166] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 775.173166] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] result = function(*args, **kwargs) [ 775.173166] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 775.173166] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] return func(*args, **kwargs) [ 775.173166] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 775.173166] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] raise e [ 775.173166] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.173166] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] nwinfo = self.network_api.allocate_for_instance( [ 775.173166] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 775.173166] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] created_port_ids = self._update_ports_for_instance( [ 775.173166] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 775.173166] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] with excutils.save_and_reraise_exception(): [ 775.173166] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.173410] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] self.force_reraise() [ 775.173410] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.173410] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] raise self.value [ 775.173410] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 775.173410] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] updated_port = self._update_port( [ 775.173410] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.173410] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] _ensure_no_port_binding_failure(port) [ 775.173410] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.173410] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] raise exception.PortBindingFailed(port_id=port['id']) [ 775.173410] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] nova.exception.PortBindingFailed: Binding failed for port f6efca39-d061-4332-a8c2-76d8cfc414b3, please check neutron logs for more information. [ 775.173410] env[61243]: ERROR nova.compute.manager [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] [ 775.173635] env[61243]: DEBUG nova.compute.utils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Binding failed for port f6efca39-d061-4332-a8c2-76d8cfc414b3, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 775.175598] env[61243]: DEBUG nova.compute.manager [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Build of instance 2e25cb9a-9b21-4b60-b741-944e53bd50e9 was re-scheduled: Binding failed for port f6efca39-d061-4332-a8c2-76d8cfc414b3, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 775.176098] env[61243]: DEBUG nova.compute.manager [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 775.176370] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Acquiring lock "refresh_cache-2e25cb9a-9b21-4b60-b741-944e53bd50e9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.176552] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Acquired lock "refresh_cache-2e25cb9a-9b21-4b60-b741-944e53bd50e9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.176743] env[61243]: DEBUG nova.network.neutron [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 775.185258] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.485s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.257243] env[61243]: DEBUG nova.network.neutron [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.423213] env[61243]: INFO nova.compute.manager [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] [instance: 454d97a2-40d0-4c7a-9df7-04c391715e0d] Took 1.02 seconds to deallocate network for instance. [ 775.698970] env[61243]: DEBUG nova.network.neutron [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.761310] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Releasing lock "refresh_cache-5314f922-6603-40dd-8760-9c1b77d6bd0a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.762667] env[61243]: DEBUG nova.compute.manager [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 775.762913] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 775.765083] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e58ad1ce-18f7-4194-bca1-287c48d14a49 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.774563] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4a86d7-6474-49c3-b410-d8ce17b26bb5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.798268] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5314f922-6603-40dd-8760-9c1b77d6bd0a could not be found. [ 775.798509] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 775.798698] env[61243]: INFO nova.compute.manager [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 775.798944] env[61243]: DEBUG oslo.service.loopingcall [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 775.801480] env[61243]: DEBUG nova.compute.manager [-] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 775.801585] env[61243]: DEBUG nova.network.neutron [-] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 775.825687] env[61243]: DEBUG nova.network.neutron [-] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.844057] env[61243]: DEBUG nova.network.neutron [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.005579] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd99097-5a6d-4b71-8291-335cabdbe838 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.013924] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d70c465-8b45-4c81-b658-fa4245096268 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.045868] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13187246-a64e-4228-b3ad-5c5a5ffacc9c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.053652] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805d8d7e-b520-40ee-8ce3-576190f4d9b6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.067301] env[61243]: DEBUG nova.compute.provider_tree [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.327924] env[61243]: DEBUG nova.network.neutron [-] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.346033] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Releasing lock "refresh_cache-2e25cb9a-9b21-4b60-b741-944e53bd50e9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.346290] env[61243]: DEBUG nova.compute.manager [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 776.346474] env[61243]: DEBUG nova.compute.manager [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 776.346645] env[61243]: DEBUG nova.network.neutron [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 776.361984] env[61243]: DEBUG nova.network.neutron [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.455097] env[61243]: INFO nova.scheduler.client.report [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Deleted allocations for instance 454d97a2-40d0-4c7a-9df7-04c391715e0d [ 776.573736] env[61243]: DEBUG nova.scheduler.client.report [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.830363] env[61243]: INFO nova.compute.manager [-] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Took 1.03 seconds to deallocate network for instance. [ 776.833901] env[61243]: DEBUG nova.compute.claims [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 776.834136] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.865361] env[61243]: DEBUG nova.network.neutron [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.965815] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bfed2c4a-12f4-4b83-8afd-1215c95b5aa1 tempest-ImagesNegativeTestJSON-1963446929 tempest-ImagesNegativeTestJSON-1963446929-project-member] Lock "454d97a2-40d0-4c7a-9df7-04c391715e0d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 167.840s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.078707] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.893s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.079344] env[61243]: ERROR nova.compute.manager [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dfcaaed7-6129-4957-b854-e8213e476069, please check neutron logs for more information. [ 777.079344] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Traceback (most recent call last): [ 777.079344] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 777.079344] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] self.driver.spawn(context, instance, image_meta, [ 777.079344] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 777.079344] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] self._vmops.spawn(context, instance, image_meta, injected_files, [ 777.079344] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 777.079344] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] vm_ref = self.build_virtual_machine(instance, [ 777.079344] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 777.079344] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] vif_infos = vmwarevif.get_vif_info(self._session, [ 777.079344] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 777.079658] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] for vif in network_info: [ 777.079658] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 777.079658] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] return self._sync_wrapper(fn, *args, **kwargs) [ 777.079658] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 777.079658] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] self.wait() [ 777.079658] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 777.079658] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] self[:] = self._gt.wait() [ 777.079658] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 777.079658] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] return self._exit_event.wait() [ 777.079658] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 777.079658] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] current.throw(*self._exc) [ 777.079658] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.079658] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] result = function(*args, **kwargs) [ 777.079976] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 777.079976] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] return func(*args, **kwargs) [ 777.079976] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.079976] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] raise e [ 777.079976] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.079976] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] nwinfo = self.network_api.allocate_for_instance( [ 777.079976] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 777.079976] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] created_port_ids = self._update_ports_for_instance( [ 777.079976] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 777.079976] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] with excutils.save_and_reraise_exception(): [ 777.079976] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.079976] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] self.force_reraise() [ 777.079976] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.080303] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] raise self.value [ 777.080303] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 777.080303] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] updated_port = self._update_port( [ 777.080303] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.080303] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] _ensure_no_port_binding_failure(port) [ 777.080303] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.080303] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] raise exception.PortBindingFailed(port_id=port['id']) [ 777.080303] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] nova.exception.PortBindingFailed: Binding failed for port dfcaaed7-6129-4957-b854-e8213e476069, please check neutron logs for more information. [ 777.080303] env[61243]: ERROR nova.compute.manager [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] [ 777.080303] env[61243]: DEBUG nova.compute.utils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Binding failed for port dfcaaed7-6129-4957-b854-e8213e476069, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 777.081387] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.991s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.085745] env[61243]: DEBUG nova.compute.manager [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Build of instance 4a0a08ca-bbf5-4124-853e-0be45f531d91 was re-scheduled: Binding failed for port dfcaaed7-6129-4957-b854-e8213e476069, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 777.086578] env[61243]: DEBUG nova.compute.manager [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 777.086578] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "refresh_cache-4a0a08ca-bbf5-4124-853e-0be45f531d91" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.086731] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "refresh_cache-4a0a08ca-bbf5-4124-853e-0be45f531d91" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.086958] env[61243]: DEBUG nova.network.neutron [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.370011] env[61243]: INFO nova.compute.manager [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] [instance: 2e25cb9a-9b21-4b60-b741-944e53bd50e9] Took 1.02 seconds to deallocate network for instance. [ 777.470962] env[61243]: DEBUG nova.compute.manager [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 777.618962] env[61243]: DEBUG nova.network.neutron [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.742554] env[61243]: DEBUG nova.network.neutron [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.995234] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.061307] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d51d6e2-4c78-42d2-b4ea-d5a7fcfcea51 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.070297] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b857aa1-3f5c-412e-a947-56b6bbca34c6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.105313] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab64862d-37e1-4ad2-aee7-c4167c537f69 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.113381] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26afc9ec-d1f6-4e17-a45d-c1faccda01e3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.127609] env[61243]: DEBUG nova.compute.provider_tree [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.245533] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "refresh_cache-4a0a08ca-bbf5-4124-853e-0be45f531d91" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.245837] env[61243]: DEBUG nova.compute.manager [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 778.246012] env[61243]: DEBUG nova.compute.manager [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.246623] env[61243]: DEBUG nova.network.neutron [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 778.266229] env[61243]: DEBUG nova.network.neutron [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.415525] env[61243]: INFO nova.scheduler.client.report [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Deleted allocations for instance 2e25cb9a-9b21-4b60-b741-944e53bd50e9 [ 778.631423] env[61243]: DEBUG nova.scheduler.client.report [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 778.768570] env[61243]: DEBUG nova.network.neutron [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.930595] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ed378764-a91c-4b17-b982-8cc7c719e214 tempest-InstanceActionsNegativeTestJSON-431759302 tempest-InstanceActionsNegativeTestJSON-431759302-project-member] Lock "2e25cb9a-9b21-4b60-b741-944e53bd50e9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 165.315s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.140223] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.058s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.140529] env[61243]: ERROR nova.compute.manager [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 708bace8-628b-46fb-a8c8-4e9364b1824e, please check neutron logs for more information. [ 779.140529] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Traceback (most recent call last): [ 779.140529] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 779.140529] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] self.driver.spawn(context, instance, image_meta, [ 779.140529] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 779.140529] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 779.140529] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 779.140529] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] vm_ref = self.build_virtual_machine(instance, [ 779.140529] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 779.140529] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] vif_infos = vmwarevif.get_vif_info(self._session, [ 779.140529] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 779.140816] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] for vif in network_info: [ 779.140816] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 779.140816] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] return self._sync_wrapper(fn, *args, **kwargs) [ 779.140816] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 779.140816] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] self.wait() [ 779.140816] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 779.140816] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] self[:] = self._gt.wait() [ 779.140816] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 779.140816] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] return self._exit_event.wait() [ 779.140816] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 779.140816] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] current.throw(*self._exc) [ 779.140816] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.140816] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] result = function(*args, **kwargs) [ 779.141123] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 779.141123] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] return func(*args, **kwargs) [ 779.141123] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.141123] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] raise e [ 779.141123] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.141123] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] nwinfo = self.network_api.allocate_for_instance( [ 779.141123] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.141123] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] created_port_ids = self._update_ports_for_instance( [ 779.141123] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.141123] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] with excutils.save_and_reraise_exception(): [ 779.141123] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.141123] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] self.force_reraise() [ 779.141123] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.141405] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] raise self.value [ 779.141405] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.141405] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] updated_port = self._update_port( [ 779.141405] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.141405] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] _ensure_no_port_binding_failure(port) [ 779.141405] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.141405] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] raise exception.PortBindingFailed(port_id=port['id']) [ 779.141405] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] nova.exception.PortBindingFailed: Binding failed for port 708bace8-628b-46fb-a8c8-4e9364b1824e, please check neutron logs for more information. [ 779.141405] env[61243]: ERROR nova.compute.manager [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] [ 779.142305] env[61243]: DEBUG nova.compute.utils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Binding failed for port 708bace8-628b-46fb-a8c8-4e9364b1824e, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 779.144727] env[61243]: DEBUG nova.compute.manager [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Build of instance 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9 was re-scheduled: Binding failed for port 708bace8-628b-46fb-a8c8-4e9364b1824e, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 779.144727] env[61243]: DEBUG nova.compute.manager [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 779.144727] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Acquiring lock "refresh_cache-0d0504ae-8a82-4b5d-9d5d-535f30ceafa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.144727] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Acquired lock "refresh_cache-0d0504ae-8a82-4b5d-9d5d-535f30ceafa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.144978] env[61243]: DEBUG nova.network.neutron [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 779.146062] env[61243]: DEBUG oslo_concurrency.lockutils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.684s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.275034] env[61243]: INFO nova.compute.manager [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 4a0a08ca-bbf5-4124-853e-0be45f531d91] Took 1.03 seconds to deallocate network for instance. [ 779.433599] env[61243]: DEBUG nova.compute.manager [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 779.671495] env[61243]: DEBUG nova.network.neutron [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.786853] env[61243]: DEBUG nova.network.neutron [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.880474] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "9a66fd05-1880-423a-8f60-4f11f1c75ab3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.880474] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "9a66fd05-1880-423a-8f60-4f11f1c75ab3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.916078] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "1dc8ad6a-3a00-47c6-8985-481e7f1363cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.916721] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "1dc8ad6a-3a00-47c6-8985-481e7f1363cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.968472] env[61243]: DEBUG oslo_concurrency.lockutils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.986800] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63068b36-98ea-43d1-ae10-ae32667ac2c5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.994978] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3035e6c2-e856-4f8b-a42b-55cf2c55b945 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.031367] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6e6c61-e4ae-40a7-9452-e9434f652f07 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.039156] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76980407-4661-4a53-8b69-3c2b19706552 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.054194] env[61243]: DEBUG nova.compute.provider_tree [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.291597] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Releasing lock "refresh_cache-0d0504ae-8a82-4b5d-9d5d-535f30ceafa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.292302] env[61243]: DEBUG nova.compute.manager [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 780.292302] env[61243]: DEBUG nova.compute.manager [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 780.292302] env[61243]: DEBUG nova.network.neutron [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 780.311057] env[61243]: DEBUG nova.network.neutron [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.316162] env[61243]: INFO nova.scheduler.client.report [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Deleted allocations for instance 4a0a08ca-bbf5-4124-853e-0be45f531d91 [ 780.557541] env[61243]: DEBUG nova.scheduler.client.report [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 780.813979] env[61243]: DEBUG nova.network.neutron [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.827711] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ce1f22df-e0af-4a5d-ab99-d019169b3106 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "4a0a08ca-bbf5-4124-853e-0be45f531d91" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 163.063s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.066544] env[61243]: DEBUG oslo_concurrency.lockutils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.920s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.069024] env[61243]: ERROR nova.compute.manager [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1912d33f-4a5c-43f1-9f74-7d3afaf17a72, please check neutron logs for more information. [ 781.069024] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Traceback (most recent call last): [ 781.069024] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 781.069024] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] self.driver.spawn(context, instance, image_meta, [ 781.069024] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 781.069024] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.069024] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.069024] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] vm_ref = self.build_virtual_machine(instance, [ 781.069024] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.069024] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.069024] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.069362] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] for vif in network_info: [ 781.069362] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 781.069362] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] return self._sync_wrapper(fn, *args, **kwargs) [ 781.069362] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 781.069362] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] self.wait() [ 781.069362] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 781.069362] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] self[:] = self._gt.wait() [ 781.069362] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.069362] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] return self._exit_event.wait() [ 781.069362] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 781.069362] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] current.throw(*self._exc) [ 781.069362] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.069362] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] result = function(*args, **kwargs) [ 781.069691] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 781.069691] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] return func(*args, **kwargs) [ 781.069691] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.069691] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] raise e [ 781.069691] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.069691] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] nwinfo = self.network_api.allocate_for_instance( [ 781.069691] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.069691] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] created_port_ids = self._update_ports_for_instance( [ 781.069691] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.069691] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] with excutils.save_and_reraise_exception(): [ 781.069691] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.069691] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] self.force_reraise() [ 781.069691] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.070106] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] raise self.value [ 781.070106] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.070106] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] updated_port = self._update_port( [ 781.070106] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.070106] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] _ensure_no_port_binding_failure(port) [ 781.070106] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.070106] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] raise exception.PortBindingFailed(port_id=port['id']) [ 781.070106] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] nova.exception.PortBindingFailed: Binding failed for port 1912d33f-4a5c-43f1-9f74-7d3afaf17a72, please check neutron logs for more information. [ 781.070106] env[61243]: ERROR nova.compute.manager [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] [ 781.070106] env[61243]: DEBUG nova.compute.utils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Binding failed for port 1912d33f-4a5c-43f1-9f74-7d3afaf17a72, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 781.074015] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.311s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.074015] env[61243]: INFO nova.compute.claims [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.075837] env[61243]: DEBUG nova.compute.manager [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Build of instance 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9 was re-scheduled: Binding failed for port 1912d33f-4a5c-43f1-9f74-7d3afaf17a72, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 781.076420] env[61243]: DEBUG nova.compute.manager [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 781.076764] env[61243]: DEBUG oslo_concurrency.lockutils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Acquiring lock "refresh_cache-2a81aa94-3e30-4f0d-b540-1f98cfa4cea9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.077011] env[61243]: DEBUG oslo_concurrency.lockutils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Acquired lock "refresh_cache-2a81aa94-3e30-4f0d-b540-1f98cfa4cea9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.077253] env[61243]: DEBUG nova.network.neutron [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 781.319713] env[61243]: INFO nova.compute.manager [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] [instance: 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9] Took 1.03 seconds to deallocate network for instance. [ 781.332136] env[61243]: DEBUG nova.compute.manager [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 781.605692] env[61243]: DEBUG nova.network.neutron [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.708667] env[61243]: DEBUG nova.network.neutron [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.859383] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.211830] env[61243]: DEBUG oslo_concurrency.lockutils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Releasing lock "refresh_cache-2a81aa94-3e30-4f0d-b540-1f98cfa4cea9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.212082] env[61243]: DEBUG nova.compute.manager [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 782.212261] env[61243]: DEBUG nova.compute.manager [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.212424] env[61243]: DEBUG nova.network.neutron [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.229786] env[61243]: DEBUG nova.network.neutron [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.351882] env[61243]: INFO nova.scheduler.client.report [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Deleted allocations for instance 0d0504ae-8a82-4b5d-9d5d-535f30ceafa9 [ 782.414124] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5388f59c-6698-4457-a021-a4b27c646b47 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.422329] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ca3141-4852-4beb-84ab-56364f201973 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.453039] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6661ce4e-a680-436f-a41e-9534ca3ff347 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.462419] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ed3aa3-fa12-4154-bca4-3a61fb1de642 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.476242] env[61243]: DEBUG nova.compute.provider_tree [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.732595] env[61243]: DEBUG nova.network.neutron [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.862856] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8b08d003-bc95-45e8-9f51-96639b800c8d tempest-ServerActionsTestJSON-385996002 tempest-ServerActionsTestJSON-385996002-project-member] Lock "0d0504ae-8a82-4b5d-9d5d-535f30ceafa9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.444s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.982024] env[61243]: DEBUG nova.scheduler.client.report [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.238019] env[61243]: INFO nova.compute.manager [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] [instance: 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9] Took 1.02 seconds to deallocate network for instance. [ 783.369302] env[61243]: DEBUG nova.compute.manager [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 783.486689] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.414s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.486689] env[61243]: DEBUG nova.compute.manager [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 783.488129] env[61243]: DEBUG oslo_concurrency.lockutils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.919s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.489810] env[61243]: INFO nova.compute.claims [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.899101] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.937581] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.937759] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.998549] env[61243]: DEBUG nova.compute.utils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 784.000200] env[61243]: DEBUG nova.compute.manager [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 784.000385] env[61243]: DEBUG nova.network.neutron [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 784.080543] env[61243]: DEBUG nova.policy [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06e247d66c8d43218302a35ead439405', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e641e72e4d2048218ef403456dbb2b82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 784.279550] env[61243]: INFO nova.scheduler.client.report [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Deleted allocations for instance 2a81aa94-3e30-4f0d-b540-1f98cfa4cea9 [ 784.508858] env[61243]: DEBUG nova.compute.manager [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 784.777056] env[61243]: DEBUG nova.network.neutron [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Successfully created port: 846ec316-a3ea-48e9-aa3c-a0abf4983e5b {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 784.790038] env[61243]: DEBUG oslo_concurrency.lockutils [None req-193ca86e-d273-4eef-b698-bda38818fb76 tempest-ServerRescueTestJSONUnderV235-1052198162 tempest-ServerRescueTestJSONUnderV235-1052198162-project-member] Lock "2a81aa94-3e30-4f0d-b540-1f98cfa4cea9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.313s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.841995] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5669e3d-074f-4a31-b60b-c349f0801fa8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.850795] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45b7cdb-ff28-4ebf-896b-d70027e010d2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.886069] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb95306c-0cdf-455b-aa04-d5cc63228db7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.894080] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46143bd3-bf97-49e5-898d-064e5417fec0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.908222] env[61243]: DEBUG nova.compute.provider_tree [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.299745] env[61243]: DEBUG nova.compute.manager [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 785.412416] env[61243]: DEBUG nova.scheduler.client.report [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.523780] env[61243]: DEBUG nova.compute.manager [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 785.557617] env[61243]: DEBUG nova.virt.hardware [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 785.557886] env[61243]: DEBUG nova.virt.hardware [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 785.558357] env[61243]: DEBUG nova.virt.hardware [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 785.558826] env[61243]: DEBUG nova.virt.hardware [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 785.559041] env[61243]: DEBUG nova.virt.hardware [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 785.559204] env[61243]: DEBUG nova.virt.hardware [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 785.559414] env[61243]: DEBUG nova.virt.hardware [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 785.559726] env[61243]: DEBUG nova.virt.hardware [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 785.559952] env[61243]: DEBUG nova.virt.hardware [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 785.560147] env[61243]: DEBUG nova.virt.hardware [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 785.560333] env[61243]: DEBUG nova.virt.hardware [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 785.562558] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c5f444-f613-42a0-8a2b-4e76780ed311 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.571623] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77723d8d-8688-45ca-ba7f-a04025177003 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.763764] env[61243]: DEBUG nova.compute.manager [req-646e06cf-b1e5-4942-afab-7712dfc5959c req-eff8fa68-dd03-41e0-b043-27915b5a4eb6 service nova] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Received event network-changed-846ec316-a3ea-48e9-aa3c-a0abf4983e5b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 785.764025] env[61243]: DEBUG nova.compute.manager [req-646e06cf-b1e5-4942-afab-7712dfc5959c req-eff8fa68-dd03-41e0-b043-27915b5a4eb6 service nova] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Refreshing instance network info cache due to event network-changed-846ec316-a3ea-48e9-aa3c-a0abf4983e5b. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 785.764249] env[61243]: DEBUG oslo_concurrency.lockutils [req-646e06cf-b1e5-4942-afab-7712dfc5959c req-eff8fa68-dd03-41e0-b043-27915b5a4eb6 service nova] Acquiring lock "refresh_cache-f42f07cd-3d94-41a3-b202-ca1fa6b75da5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.764389] env[61243]: DEBUG oslo_concurrency.lockutils [req-646e06cf-b1e5-4942-afab-7712dfc5959c req-eff8fa68-dd03-41e0-b043-27915b5a4eb6 service nova] Acquired lock "refresh_cache-f42f07cd-3d94-41a3-b202-ca1fa6b75da5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.764544] env[61243]: DEBUG nova.network.neutron [req-646e06cf-b1e5-4942-afab-7712dfc5959c req-eff8fa68-dd03-41e0-b043-27915b5a4eb6 service nova] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Refreshing network info cache for port 846ec316-a3ea-48e9-aa3c-a0abf4983e5b {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 785.827548] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.839113] env[61243]: ERROR nova.compute.manager [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 846ec316-a3ea-48e9-aa3c-a0abf4983e5b, please check neutron logs for more information. [ 785.839113] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 785.839113] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 785.839113] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 785.839113] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 785.839113] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 785.839113] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 785.839113] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 785.839113] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 785.839113] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 785.839113] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 785.839113] env[61243]: ERROR nova.compute.manager raise self.value [ 785.839113] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 785.839113] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 785.839113] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 785.839113] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 785.839494] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 785.839494] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 785.839494] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 846ec316-a3ea-48e9-aa3c-a0abf4983e5b, please check neutron logs for more information. [ 785.839494] env[61243]: ERROR nova.compute.manager [ 785.839494] env[61243]: Traceback (most recent call last): [ 785.839494] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 785.839494] env[61243]: listener.cb(fileno) [ 785.839494] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 785.839494] env[61243]: result = function(*args, **kwargs) [ 785.839494] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 785.839494] env[61243]: return func(*args, **kwargs) [ 785.839494] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 785.839494] env[61243]: raise e [ 785.839494] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 785.839494] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 785.839494] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 785.839494] env[61243]: created_port_ids = self._update_ports_for_instance( [ 785.839494] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 785.839494] env[61243]: with excutils.save_and_reraise_exception(): [ 785.839494] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 785.839494] env[61243]: self.force_reraise() [ 785.839494] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 785.839494] env[61243]: raise self.value [ 785.839494] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 785.839494] env[61243]: updated_port = self._update_port( [ 785.839494] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 785.839494] env[61243]: _ensure_no_port_binding_failure(port) [ 785.839494] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 785.839494] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 785.840786] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 846ec316-a3ea-48e9-aa3c-a0abf4983e5b, please check neutron logs for more information. [ 785.840786] env[61243]: Removing descriptor: 16 [ 785.840786] env[61243]: ERROR nova.compute.manager [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 846ec316-a3ea-48e9-aa3c-a0abf4983e5b, please check neutron logs for more information. [ 785.840786] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Traceback (most recent call last): [ 785.840786] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 785.840786] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] yield resources [ 785.840786] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 785.840786] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] self.driver.spawn(context, instance, image_meta, [ 785.840786] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 785.840786] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 785.840786] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 785.840786] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] vm_ref = self.build_virtual_machine(instance, [ 785.841081] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 785.841081] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] vif_infos = vmwarevif.get_vif_info(self._session, [ 785.841081] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 785.841081] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] for vif in network_info: [ 785.841081] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 785.841081] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] return self._sync_wrapper(fn, *args, **kwargs) [ 785.841081] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 785.841081] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] self.wait() [ 785.841081] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 785.841081] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] self[:] = self._gt.wait() [ 785.841081] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 785.841081] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] return self._exit_event.wait() [ 785.841081] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 785.841354] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] result = hub.switch() [ 785.841354] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 785.841354] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] return self.greenlet.switch() [ 785.841354] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 785.841354] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] result = function(*args, **kwargs) [ 785.841354] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 785.841354] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] return func(*args, **kwargs) [ 785.841354] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 785.841354] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] raise e [ 785.841354] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 785.841354] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] nwinfo = self.network_api.allocate_for_instance( [ 785.841354] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 785.841354] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] created_port_ids = self._update_ports_for_instance( [ 785.841661] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 785.841661] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] with excutils.save_and_reraise_exception(): [ 785.841661] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 785.841661] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] self.force_reraise() [ 785.841661] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 785.841661] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] raise self.value [ 785.841661] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 785.841661] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] updated_port = self._update_port( [ 785.841661] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 785.841661] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] _ensure_no_port_binding_failure(port) [ 785.841661] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 785.841661] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] raise exception.PortBindingFailed(port_id=port['id']) [ 785.841975] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] nova.exception.PortBindingFailed: Binding failed for port 846ec316-a3ea-48e9-aa3c-a0abf4983e5b, please check neutron logs for more information. [ 785.841975] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] [ 785.841975] env[61243]: INFO nova.compute.manager [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Terminating instance [ 785.846126] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "refresh_cache-f42f07cd-3d94-41a3-b202-ca1fa6b75da5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.917708] env[61243]: DEBUG oslo_concurrency.lockutils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.918243] env[61243]: DEBUG nova.compute.manager [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.921254] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.475s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.299434] env[61243]: DEBUG nova.network.neutron [req-646e06cf-b1e5-4942-afab-7712dfc5959c req-eff8fa68-dd03-41e0-b043-27915b5a4eb6 service nova] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.431562] env[61243]: DEBUG nova.compute.utils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.437209] env[61243]: DEBUG nova.compute.manager [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 786.437209] env[61243]: DEBUG nova.network.neutron [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 786.500251] env[61243]: DEBUG nova.network.neutron [req-646e06cf-b1e5-4942-afab-7712dfc5959c req-eff8fa68-dd03-41e0-b043-27915b5a4eb6 service nova] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.541283] env[61243]: DEBUG nova.policy [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '28d7b73f55f743a1a7126cc85613ebb7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d983b7bb0224a50a26b0334e697422f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 786.939712] env[61243]: DEBUG nova.compute.manager [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 786.963655] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 1f4985a7-f6cc-4e79-aea3-ce0623aed515 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 786.963655] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 5314f922-6603-40dd-8760-9c1b77d6bd0a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 786.964391] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance f42f07cd-3d94-41a3-b202-ca1fa6b75da5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 786.964620] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 6bd8ea50-e23f-48e7-9979-019e7a694c66 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.005165] env[61243]: DEBUG oslo_concurrency.lockutils [req-646e06cf-b1e5-4942-afab-7712dfc5959c req-eff8fa68-dd03-41e0-b043-27915b5a4eb6 service nova] Releasing lock "refresh_cache-f42f07cd-3d94-41a3-b202-ca1fa6b75da5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.005165] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "refresh_cache-f42f07cd-3d94-41a3-b202-ca1fa6b75da5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.005165] env[61243]: DEBUG nova.network.neutron [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 787.075268] env[61243]: DEBUG nova.network.neutron [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Successfully created port: 047ef0cd-631f-4b18-8690-72576798bb0a {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 787.145554] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Acquiring lock "36db1ed5-846f-4ad6-8cee-38b73ff00321" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.145791] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Lock "36db1ed5-846f-4ad6-8cee-38b73ff00321" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.468268] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance e9b2e5d1-e75e-4346-a0af-c0dafac0692e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 787.634063] env[61243]: DEBUG nova.network.neutron [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.799027] env[61243]: DEBUG nova.network.neutron [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.803167] env[61243]: DEBUG nova.compute.manager [req-95cb0f41-970d-4b30-b464-22dc040ba823 req-4cbf34fc-e3f5-4429-888f-6c8fc0dca888 service nova] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Received event network-vif-deleted-846ec316-a3ea-48e9-aa3c-a0abf4983e5b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.864406] env[61243]: DEBUG oslo_concurrency.lockutils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.864406] env[61243]: DEBUG oslo_concurrency.lockutils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.959259] env[61243]: DEBUG nova.compute.manager [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 787.974393] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 86179fb0-99df-4b10-a815-c19168e9521e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 787.997578] env[61243]: DEBUG nova.virt.hardware [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 787.997578] env[61243]: DEBUG nova.virt.hardware [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 787.997710] env[61243]: DEBUG nova.virt.hardware [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.997893] env[61243]: DEBUG nova.virt.hardware [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 787.998427] env[61243]: DEBUG nova.virt.hardware [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.998427] env[61243]: DEBUG nova.virt.hardware [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 787.998652] env[61243]: DEBUG nova.virt.hardware [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 787.998834] env[61243]: DEBUG nova.virt.hardware [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 787.999124] env[61243]: DEBUG nova.virt.hardware [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 787.999217] env[61243]: DEBUG nova.virt.hardware [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 787.999350] env[61243]: DEBUG nova.virt.hardware [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.000675] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea8f81e9-9073-4396-a3dd-7e707ed5b973 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.010135] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31107dc6-e4f9-40a0-a0f9-3867808c9b7c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.300720] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "refresh_cache-f42f07cd-3d94-41a3-b202-ca1fa6b75da5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.301168] env[61243]: DEBUG nova.compute.manager [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 788.301365] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 788.301686] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5d9a18d4-9acd-470d-835f-b62723c97336 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.310782] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d8e65e7-5598-467f-a465-b4f21280c338 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.333583] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f42f07cd-3d94-41a3-b202-ca1fa6b75da5 could not be found. [ 788.333816] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 788.334047] env[61243]: INFO nova.compute.manager [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 788.334325] env[61243]: DEBUG oslo.service.loopingcall [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 788.334554] env[61243]: DEBUG nova.compute.manager [-] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 788.334645] env[61243]: DEBUG nova.network.neutron [-] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 788.364211] env[61243]: DEBUG nova.network.neutron [-] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.478816] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 07ee984d-476e-484d-ba80-0ec2e411faa9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 788.865533] env[61243]: DEBUG nova.network.neutron [-] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.959747] env[61243]: ERROR nova.compute.manager [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 047ef0cd-631f-4b18-8690-72576798bb0a, please check neutron logs for more information. [ 788.959747] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 788.959747] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 788.959747] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 788.959747] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 788.959747] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 788.959747] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 788.959747] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 788.959747] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 788.959747] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 788.959747] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 788.959747] env[61243]: ERROR nova.compute.manager raise self.value [ 788.959747] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 788.959747] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 788.959747] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 788.959747] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 788.960433] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 788.960433] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 788.960433] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 047ef0cd-631f-4b18-8690-72576798bb0a, please check neutron logs for more information. [ 788.960433] env[61243]: ERROR nova.compute.manager [ 788.960433] env[61243]: Traceback (most recent call last): [ 788.960433] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 788.960433] env[61243]: listener.cb(fileno) [ 788.960433] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 788.960433] env[61243]: result = function(*args, **kwargs) [ 788.960433] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 788.960433] env[61243]: return func(*args, **kwargs) [ 788.960433] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 788.960433] env[61243]: raise e [ 788.960433] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 788.960433] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 788.960433] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 788.960433] env[61243]: created_port_ids = self._update_ports_for_instance( [ 788.960433] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 788.960433] env[61243]: with excutils.save_and_reraise_exception(): [ 788.960433] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 788.960433] env[61243]: self.force_reraise() [ 788.960433] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 788.960433] env[61243]: raise self.value [ 788.960433] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 788.960433] env[61243]: updated_port = self._update_port( [ 788.960433] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 788.960433] env[61243]: _ensure_no_port_binding_failure(port) [ 788.960433] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 788.960433] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 788.961158] env[61243]: nova.exception.PortBindingFailed: Binding failed for port 047ef0cd-631f-4b18-8690-72576798bb0a, please check neutron logs for more information. [ 788.961158] env[61243]: Removing descriptor: 15 [ 788.961465] env[61243]: ERROR nova.compute.manager [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 047ef0cd-631f-4b18-8690-72576798bb0a, please check neutron logs for more information. [ 788.961465] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Traceback (most recent call last): [ 788.961465] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 788.961465] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] yield resources [ 788.961465] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 788.961465] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] self.driver.spawn(context, instance, image_meta, [ 788.961465] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 788.961465] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] self._vmops.spawn(context, instance, image_meta, injected_files, [ 788.961465] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 788.961465] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] vm_ref = self.build_virtual_machine(instance, [ 788.961465] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 788.961731] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] vif_infos = vmwarevif.get_vif_info(self._session, [ 788.961731] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 788.961731] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] for vif in network_info: [ 788.961731] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 788.961731] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] return self._sync_wrapper(fn, *args, **kwargs) [ 788.961731] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 788.961731] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] self.wait() [ 788.961731] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 788.961731] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] self[:] = self._gt.wait() [ 788.961731] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 788.961731] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] return self._exit_event.wait() [ 788.961731] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 788.961731] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] result = hub.switch() [ 788.962086] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 788.962086] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] return self.greenlet.switch() [ 788.962086] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 788.962086] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] result = function(*args, **kwargs) [ 788.962086] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 788.962086] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] return func(*args, **kwargs) [ 788.962086] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 788.962086] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] raise e [ 788.962086] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 788.962086] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] nwinfo = self.network_api.allocate_for_instance( [ 788.962086] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 788.962086] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] created_port_ids = self._update_ports_for_instance( [ 788.962086] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 788.962418] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] with excutils.save_and_reraise_exception(): [ 788.962418] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 788.962418] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] self.force_reraise() [ 788.962418] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 788.962418] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] raise self.value [ 788.962418] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 788.962418] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] updated_port = self._update_port( [ 788.962418] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 788.962418] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] _ensure_no_port_binding_failure(port) [ 788.962418] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 788.962418] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] raise exception.PortBindingFailed(port_id=port['id']) [ 788.962418] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] nova.exception.PortBindingFailed: Binding failed for port 047ef0cd-631f-4b18-8690-72576798bb0a, please check neutron logs for more information. [ 788.962418] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] [ 788.962754] env[61243]: INFO nova.compute.manager [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Terminating instance [ 788.964247] env[61243]: DEBUG oslo_concurrency.lockutils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Acquiring lock "refresh_cache-6bd8ea50-e23f-48e7-9979-019e7a694c66" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.964455] env[61243]: DEBUG oslo_concurrency.lockutils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Acquired lock "refresh_cache-6bd8ea50-e23f-48e7-9979-019e7a694c66" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.964655] env[61243]: DEBUG nova.network.neutron [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 788.982290] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 36e55334-8628-4dd7-a845-f4ae3d8e7ff9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.369223] env[61243]: INFO nova.compute.manager [-] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Took 1.03 seconds to deallocate network for instance. [ 789.371581] env[61243]: DEBUG nova.compute.claims [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 789.371774] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.481484] env[61243]: DEBUG nova.network.neutron [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.486068] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.562667] env[61243]: DEBUG nova.network.neutron [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.839588] env[61243]: DEBUG nova.compute.manager [req-4d72b012-5d5a-4eff-a040-38563509b958 req-44dc0da1-d70c-484e-b4ae-be4549b7b2e7 service nova] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Received event network-changed-047ef0cd-631f-4b18-8690-72576798bb0a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.839989] env[61243]: DEBUG nova.compute.manager [req-4d72b012-5d5a-4eff-a040-38563509b958 req-44dc0da1-d70c-484e-b4ae-be4549b7b2e7 service nova] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Refreshing instance network info cache due to event network-changed-047ef0cd-631f-4b18-8690-72576798bb0a. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 789.840369] env[61243]: DEBUG oslo_concurrency.lockutils [req-4d72b012-5d5a-4eff-a040-38563509b958 req-44dc0da1-d70c-484e-b4ae-be4549b7b2e7 service nova] Acquiring lock "refresh_cache-6bd8ea50-e23f-48e7-9979-019e7a694c66" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.992098] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 6cee3dd1-c13d-427d-889c-6cd3ae388a31 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 790.066434] env[61243]: DEBUG oslo_concurrency.lockutils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Releasing lock "refresh_cache-6bd8ea50-e23f-48e7-9979-019e7a694c66" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.066919] env[61243]: DEBUG nova.compute.manager [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 790.067187] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 790.067547] env[61243]: DEBUG oslo_concurrency.lockutils [req-4d72b012-5d5a-4eff-a040-38563509b958 req-44dc0da1-d70c-484e-b4ae-be4549b7b2e7 service nova] Acquired lock "refresh_cache-6bd8ea50-e23f-48e7-9979-019e7a694c66" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.067769] env[61243]: DEBUG nova.network.neutron [req-4d72b012-5d5a-4eff-a040-38563509b958 req-44dc0da1-d70c-484e-b4ae-be4549b7b2e7 service nova] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Refreshing network info cache for port 047ef0cd-631f-4b18-8690-72576798bb0a {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 790.069223] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-97b611b5-ebb9-47f2-9c04-20a0efaefee9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.078903] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f149a4e-04a8-4c2f-8347-7433e26bfb3f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.102739] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6bd8ea50-e23f-48e7-9979-019e7a694c66 could not be found. [ 790.103083] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 790.103802] env[61243]: INFO nova.compute.manager [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Took 0.04 seconds to destroy the instance on the hypervisor. [ 790.103802] env[61243]: DEBUG oslo.service.loopingcall [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.103802] env[61243]: DEBUG nova.compute.manager [-] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 790.103802] env[61243]: DEBUG nova.network.neutron [-] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 790.125850] env[61243]: DEBUG nova.network.neutron [-] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.495844] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 894eb798-6c9e-47cb-8eb5-4610fe184bc5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 790.591088] env[61243]: DEBUG nova.network.neutron [req-4d72b012-5d5a-4eff-a040-38563509b958 req-44dc0da1-d70c-484e-b4ae-be4549b7b2e7 service nova] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.628382] env[61243]: DEBUG nova.network.neutron [-] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.706132] env[61243]: DEBUG nova.network.neutron [req-4d72b012-5d5a-4eff-a040-38563509b958 req-44dc0da1-d70c-484e-b4ae-be4549b7b2e7 service nova] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.001380] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 791.136125] env[61243]: INFO nova.compute.manager [-] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Took 1.03 seconds to deallocate network for instance. [ 791.137088] env[61243]: DEBUG nova.compute.claims [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 791.137088] env[61243]: DEBUG oslo_concurrency.lockutils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.209543] env[61243]: DEBUG oslo_concurrency.lockutils [req-4d72b012-5d5a-4eff-a040-38563509b958 req-44dc0da1-d70c-484e-b4ae-be4549b7b2e7 service nova] Releasing lock "refresh_cache-6bd8ea50-e23f-48e7-9979-019e7a694c66" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.209543] env[61243]: DEBUG nova.compute.manager [req-4d72b012-5d5a-4eff-a040-38563509b958 req-44dc0da1-d70c-484e-b4ae-be4549b7b2e7 service nova] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Received event network-vif-deleted-047ef0cd-631f-4b18-8690-72576798bb0a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.505883] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 729d9bd9-1ab0-47bd-9b7a-11412d3dc608 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 791.893184] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "c1a0d242-4eae-4c03-8341-840b41341f17" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.893481] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "c1a0d242-4eae-4c03-8341-840b41341f17" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.007966] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 96315d9e-4eda-4e3a-af0d-bdc52ab181e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 792.511738] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 6f33b303-1ecd-478e-8fa3-33a5f9a5af50 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 793.014581] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 9a66fd05-1880-423a-8f60-4f11f1c75ab3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 793.517207] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 1dc8ad6a-3a00-47c6-8985-481e7f1363cd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 794.020600] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 794.525321] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 36db1ed5-846f-4ad6-8cee-38b73ff00321 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 794.525321] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 794.525321] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 794.751130] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8225788-dbb8-4519-8a78-a22617f82df9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.759551] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c885f711-c529-43b1-9d27-849d91582a3e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.788518] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e03ffa9-9697-46a5-90a6-bd2cd3010aa6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.796708] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9814ca56-948c-4e23-bfef-230b8ee3490e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.810240] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.313678] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.819501] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61243) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 795.819760] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 9.899s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.820056] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.887s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.822829] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 795.822992] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Cleaning up deleted instances {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 796.329990] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] There are 3 instances to clean {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 796.330316] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: fd2d724a-76b6-405f-b003-54de2ad84f5a] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 796.537625] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe3bbcd-fbb2-4354-8528-79f31abf2bf6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.544681] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96715269-0c6c-4a48-9698-f6f70920ff94 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.573927] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f3b790-6208-4880-80a4-a24aaa9b05a4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.583646] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b21c13-1e11-4259-a6da-26b61cfa9c20 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.597032] env[61243]: DEBUG nova.compute.provider_tree [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.833414] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: b5163d89-ceb8-4c61-ae02-0ae3311b58a9] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 797.100464] env[61243]: DEBUG nova.scheduler.client.report [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.337617] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 1b1b2bac-67af-4921-a49a-429017ef7a3b] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 797.605534] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.785s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.606225] env[61243]: ERROR nova.compute.manager [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0bcb6106-50de-4d90-80ac-ce28ca91123a, please check neutron logs for more information. [ 797.606225] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Traceback (most recent call last): [ 797.606225] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 797.606225] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] self.driver.spawn(context, instance, image_meta, [ 797.606225] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 797.606225] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] self._vmops.spawn(context, instance, image_meta, injected_files, [ 797.606225] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 797.606225] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] vm_ref = self.build_virtual_machine(instance, [ 797.606225] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 797.606225] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] vif_infos = vmwarevif.get_vif_info(self._session, [ 797.606225] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 797.606587] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] for vif in network_info: [ 797.606587] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 797.606587] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] return self._sync_wrapper(fn, *args, **kwargs) [ 797.606587] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 797.606587] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] self.wait() [ 797.606587] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 797.606587] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] self[:] = self._gt.wait() [ 797.606587] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 797.606587] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] return self._exit_event.wait() [ 797.606587] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 797.606587] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] result = hub.switch() [ 797.606587] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 797.606587] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] return self.greenlet.switch() [ 797.607075] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 797.607075] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] result = function(*args, **kwargs) [ 797.607075] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 797.607075] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] return func(*args, **kwargs) [ 797.607075] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 797.607075] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] raise e [ 797.607075] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 797.607075] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] nwinfo = self.network_api.allocate_for_instance( [ 797.607075] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 797.607075] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] created_port_ids = self._update_ports_for_instance( [ 797.607075] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 797.607075] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] with excutils.save_and_reraise_exception(): [ 797.607075] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 797.607480] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] self.force_reraise() [ 797.607480] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 797.607480] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] raise self.value [ 797.607480] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 797.607480] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] updated_port = self._update_port( [ 797.607480] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 797.607480] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] _ensure_no_port_binding_failure(port) [ 797.607480] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 797.607480] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] raise exception.PortBindingFailed(port_id=port['id']) [ 797.607480] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] nova.exception.PortBindingFailed: Binding failed for port 0bcb6106-50de-4d90-80ac-ce28ca91123a, please check neutron logs for more information. [ 797.607480] env[61243]: ERROR nova.compute.manager [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] [ 797.607793] env[61243]: DEBUG nova.compute.utils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Binding failed for port 0bcb6106-50de-4d90-80ac-ce28ca91123a, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 797.608512] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.767s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.609930] env[61243]: INFO nova.compute.claims [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 797.612458] env[61243]: DEBUG nova.compute.manager [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Build of instance 1f4985a7-f6cc-4e79-aea3-ce0623aed515 was re-scheduled: Binding failed for port 0bcb6106-50de-4d90-80ac-ce28ca91123a, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 797.612888] env[61243]: DEBUG nova.compute.manager [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 797.613128] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Acquiring lock "refresh_cache-1f4985a7-f6cc-4e79-aea3-ce0623aed515" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.613282] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Acquired lock "refresh_cache-1f4985a7-f6cc-4e79-aea3-ce0623aed515" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.613442] env[61243]: DEBUG nova.network.neutron [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 797.841619] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 797.841792] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Cleaning up deleted instances with incomplete migration {{(pid=61243) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 798.134553] env[61243]: DEBUG nova.network.neutron [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.208867] env[61243]: DEBUG nova.network.neutron [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.344242] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 799.182231] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Releasing lock "refresh_cache-1f4985a7-f6cc-4e79-aea3-ce0623aed515" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.182486] env[61243]: DEBUG nova.compute.manager [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 799.182657] env[61243]: DEBUG nova.compute.manager [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 799.182819] env[61243]: DEBUG nova.network.neutron [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 799.198783] env[61243]: DEBUG nova.network.neutron [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 799.410136] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e1c133-6c51-4271-bf10-33e5a1e99502 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.419168] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab30d6a-58b2-471c-93f2-e82859eb6709 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.446383] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22306bf4-37a8-4411-b0ea-711916803083 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.454972] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0793092d-24ea-412e-91aa-39f21116da20 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.466133] env[61243]: DEBUG nova.compute.provider_tree [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.701036] env[61243]: DEBUG nova.network.neutron [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.969231] env[61243]: DEBUG nova.scheduler.client.report [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.204573] env[61243]: INFO nova.compute.manager [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] [instance: 1f4985a7-f6cc-4e79-aea3-ce0623aed515] Took 1.02 seconds to deallocate network for instance. [ 800.474186] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.866s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.474775] env[61243]: DEBUG nova.compute.manager [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 800.478356] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.643s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.983045] env[61243]: DEBUG nova.compute.utils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.988123] env[61243]: DEBUG nova.compute.manager [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 800.988334] env[61243]: DEBUG nova.network.neutron [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 801.050314] env[61243]: DEBUG nova.policy [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '82364638eb3a41e7a53ede692e85411b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c311cfe8e8b42908ff1931150e65701', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 801.232986] env[61243]: INFO nova.scheduler.client.report [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Deleted allocations for instance 1f4985a7-f6cc-4e79-aea3-ce0623aed515 [ 801.243911] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb69804-2b69-416d-9730-63224ae80856 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.252311] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff29ce77-9afb-4325-a699-a1bd6b8b5b67 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.287257] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b6903b-badf-43b1-9e25-bae977146fcb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.293618] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-822fb4f0-d8b2-444a-8841-6a031a1de577 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.307690] env[61243]: DEBUG nova.compute.provider_tree [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.334245] env[61243]: DEBUG nova.network.neutron [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Successfully created port: cff38ea8-81bd-4bd5-b704-b61286d84199 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.488753] env[61243]: DEBUG nova.compute.manager [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 801.743785] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f3b6a05-9a4e-48ce-814c-816a15655742 tempest-ServersTestMultiNic-1879496255 tempest-ServersTestMultiNic-1879496255-project-member] Lock "1f4985a7-f6cc-4e79-aea3-ce0623aed515" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.476s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.810375] env[61243]: DEBUG nova.scheduler.client.report [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.033484] env[61243]: DEBUG nova.compute.manager [req-501104be-8de7-49f9-a0d8-e55dc598e1b5 req-8ecf95f6-f27c-4e71-ab03-50d5f0e3da41 service nova] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Received event network-changed-cff38ea8-81bd-4bd5-b704-b61286d84199 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.033484] env[61243]: DEBUG nova.compute.manager [req-501104be-8de7-49f9-a0d8-e55dc598e1b5 req-8ecf95f6-f27c-4e71-ab03-50d5f0e3da41 service nova] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Refreshing instance network info cache due to event network-changed-cff38ea8-81bd-4bd5-b704-b61286d84199. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 802.033484] env[61243]: DEBUG oslo_concurrency.lockutils [req-501104be-8de7-49f9-a0d8-e55dc598e1b5 req-8ecf95f6-f27c-4e71-ab03-50d5f0e3da41 service nova] Acquiring lock "refresh_cache-e9b2e5d1-e75e-4346-a0af-c0dafac0692e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.033484] env[61243]: DEBUG oslo_concurrency.lockutils [req-501104be-8de7-49f9-a0d8-e55dc598e1b5 req-8ecf95f6-f27c-4e71-ab03-50d5f0e3da41 service nova] Acquired lock "refresh_cache-e9b2e5d1-e75e-4346-a0af-c0dafac0692e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.033484] env[61243]: DEBUG nova.network.neutron [req-501104be-8de7-49f9-a0d8-e55dc598e1b5 req-8ecf95f6-f27c-4e71-ab03-50d5f0e3da41 service nova] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Refreshing network info cache for port cff38ea8-81bd-4bd5-b704-b61286d84199 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 802.216835] env[61243]: ERROR nova.compute.manager [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cff38ea8-81bd-4bd5-b704-b61286d84199, please check neutron logs for more information. [ 802.216835] env[61243]: ERROR nova.compute.manager Traceback (most recent call last): [ 802.216835] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 802.216835] env[61243]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 802.216835] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 802.216835] env[61243]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 802.216835] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 802.216835] env[61243]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 802.216835] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 802.216835] env[61243]: ERROR nova.compute.manager self.force_reraise() [ 802.216835] env[61243]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 802.216835] env[61243]: ERROR nova.compute.manager raise self.value [ 802.216835] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 802.216835] env[61243]: ERROR nova.compute.manager updated_port = self._update_port( [ 802.216835] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 802.216835] env[61243]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 802.217305] env[61243]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 802.217305] env[61243]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 802.217305] env[61243]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cff38ea8-81bd-4bd5-b704-b61286d84199, please check neutron logs for more information. [ 802.217305] env[61243]: ERROR nova.compute.manager [ 802.217305] env[61243]: Traceback (most recent call last): [ 802.217305] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 802.217305] env[61243]: listener.cb(fileno) [ 802.217305] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 802.217305] env[61243]: result = function(*args, **kwargs) [ 802.217305] env[61243]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 802.217305] env[61243]: return func(*args, **kwargs) [ 802.217305] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 802.217305] env[61243]: raise e [ 802.217305] env[61243]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 802.217305] env[61243]: nwinfo = self.network_api.allocate_for_instance( [ 802.217305] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 802.217305] env[61243]: created_port_ids = self._update_ports_for_instance( [ 802.217305] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 802.217305] env[61243]: with excutils.save_and_reraise_exception(): [ 802.217305] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 802.217305] env[61243]: self.force_reraise() [ 802.217305] env[61243]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 802.217305] env[61243]: raise self.value [ 802.217305] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 802.217305] env[61243]: updated_port = self._update_port( [ 802.217305] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 802.217305] env[61243]: _ensure_no_port_binding_failure(port) [ 802.217305] env[61243]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 802.217305] env[61243]: raise exception.PortBindingFailed(port_id=port['id']) [ 802.217986] env[61243]: nova.exception.PortBindingFailed: Binding failed for port cff38ea8-81bd-4bd5-b704-b61286d84199, please check neutron logs for more information. [ 802.217986] env[61243]: Removing descriptor: 15 [ 802.246486] env[61243]: DEBUG nova.compute.manager [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 802.315707] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.838s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.316439] env[61243]: ERROR nova.compute.manager [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2c60b335-f40e-41b2-9f29-07d59e2562c6, please check neutron logs for more information. [ 802.316439] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Traceback (most recent call last): [ 802.316439] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 802.316439] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] self.driver.spawn(context, instance, image_meta, [ 802.316439] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 802.316439] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 802.316439] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 802.316439] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] vm_ref = self.build_virtual_machine(instance, [ 802.316439] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 802.316439] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] vif_infos = vmwarevif.get_vif_info(self._session, [ 802.316439] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 802.316899] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] for vif in network_info: [ 802.316899] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 802.316899] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] return self._sync_wrapper(fn, *args, **kwargs) [ 802.316899] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 802.316899] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] self.wait() [ 802.316899] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 802.316899] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] self[:] = self._gt.wait() [ 802.316899] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 802.316899] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] return self._exit_event.wait() [ 802.316899] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 802.316899] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] current.throw(*self._exc) [ 802.316899] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 802.316899] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] result = function(*args, **kwargs) [ 802.317876] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 802.317876] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] return func(*args, **kwargs) [ 802.317876] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 802.317876] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] raise e [ 802.317876] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 802.317876] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] nwinfo = self.network_api.allocate_for_instance( [ 802.317876] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 802.317876] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] created_port_ids = self._update_ports_for_instance( [ 802.317876] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 802.317876] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] with excutils.save_and_reraise_exception(): [ 802.317876] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 802.317876] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] self.force_reraise() [ 802.317876] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 802.318699] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] raise self.value [ 802.318699] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 802.318699] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] updated_port = self._update_port( [ 802.318699] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 802.318699] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] _ensure_no_port_binding_failure(port) [ 802.318699] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 802.318699] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] raise exception.PortBindingFailed(port_id=port['id']) [ 802.318699] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] nova.exception.PortBindingFailed: Binding failed for port 2c60b335-f40e-41b2-9f29-07d59e2562c6, please check neutron logs for more information. [ 802.318699] env[61243]: ERROR nova.compute.manager [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] [ 802.318699] env[61243]: DEBUG nova.compute.utils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Binding failed for port 2c60b335-f40e-41b2-9f29-07d59e2562c6, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 802.319360] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.324s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.321237] env[61243]: INFO nova.compute.claims [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 802.323873] env[61243]: DEBUG nova.compute.manager [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Build of instance 5314f922-6603-40dd-8760-9c1b77d6bd0a was re-scheduled: Binding failed for port 2c60b335-f40e-41b2-9f29-07d59e2562c6, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 802.324461] env[61243]: DEBUG nova.compute.manager [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 802.324670] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Acquiring lock "refresh_cache-5314f922-6603-40dd-8760-9c1b77d6bd0a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.324962] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Acquired lock "refresh_cache-5314f922-6603-40dd-8760-9c1b77d6bd0a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.324962] env[61243]: DEBUG nova.network.neutron [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 802.500123] env[61243]: DEBUG nova.compute.manager [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 802.527789] env[61243]: DEBUG nova.virt.hardware [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.528060] env[61243]: DEBUG nova.virt.hardware [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.528234] env[61243]: DEBUG nova.virt.hardware [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.528401] env[61243]: DEBUG nova.virt.hardware [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.528547] env[61243]: DEBUG nova.virt.hardware [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.528693] env[61243]: DEBUG nova.virt.hardware [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.528895] env[61243]: DEBUG nova.virt.hardware [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.529065] env[61243]: DEBUG nova.virt.hardware [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.529234] env[61243]: DEBUG nova.virt.hardware [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.529393] env[61243]: DEBUG nova.virt.hardware [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.529563] env[61243]: DEBUG nova.virt.hardware [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.530439] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4454cfe-2f26-4a3c-9b60-7a15521541fe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.540760] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d2f8a2-9fb0-4cc1-b1ac-13193ed1e225 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.554610] env[61243]: ERROR nova.compute.manager [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cff38ea8-81bd-4bd5-b704-b61286d84199, please check neutron logs for more information. [ 802.554610] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Traceback (most recent call last): [ 802.554610] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 802.554610] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] yield resources [ 802.554610] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 802.554610] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] self.driver.spawn(context, instance, image_meta, [ 802.554610] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 802.554610] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 802.554610] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 802.554610] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] vm_ref = self.build_virtual_machine(instance, [ 802.554610] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 802.555065] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] vif_infos = vmwarevif.get_vif_info(self._session, [ 802.555065] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 802.555065] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] for vif in network_info: [ 802.555065] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 802.555065] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] return self._sync_wrapper(fn, *args, **kwargs) [ 802.555065] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 802.555065] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] self.wait() [ 802.555065] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 802.555065] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] self[:] = self._gt.wait() [ 802.555065] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 802.555065] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] return self._exit_event.wait() [ 802.555065] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 802.555065] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] current.throw(*self._exc) [ 802.555368] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 802.555368] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] result = function(*args, **kwargs) [ 802.555368] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 802.555368] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] return func(*args, **kwargs) [ 802.555368] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 802.555368] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] raise e [ 802.555368] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 802.555368] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] nwinfo = self.network_api.allocate_for_instance( [ 802.555368] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 802.555368] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] created_port_ids = self._update_ports_for_instance( [ 802.555368] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 802.555368] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] with excutils.save_and_reraise_exception(): [ 802.555368] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 802.555714] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] self.force_reraise() [ 802.555714] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 802.555714] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] raise self.value [ 802.555714] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 802.555714] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] updated_port = self._update_port( [ 802.555714] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 802.555714] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] _ensure_no_port_binding_failure(port) [ 802.555714] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 802.555714] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] raise exception.PortBindingFailed(port_id=port['id']) [ 802.555714] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] nova.exception.PortBindingFailed: Binding failed for port cff38ea8-81bd-4bd5-b704-b61286d84199, please check neutron logs for more information. [ 802.555714] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] [ 802.555714] env[61243]: INFO nova.compute.manager [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Terminating instance [ 802.557081] env[61243]: DEBUG nova.network.neutron [req-501104be-8de7-49f9-a0d8-e55dc598e1b5 req-8ecf95f6-f27c-4e71-ab03-50d5f0e3da41 service nova] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.559039] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Acquiring lock "refresh_cache-e9b2e5d1-e75e-4346-a0af-c0dafac0692e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.671869] env[61243]: DEBUG nova.network.neutron [req-501104be-8de7-49f9-a0d8-e55dc598e1b5 req-8ecf95f6-f27c-4e71-ab03-50d5f0e3da41 service nova] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.767847] env[61243]: DEBUG oslo_concurrency.lockutils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.848988] env[61243]: DEBUG nova.network.neutron [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 803.092399] env[61243]: DEBUG nova.network.neutron [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.173662] env[61243]: DEBUG oslo_concurrency.lockutils [req-501104be-8de7-49f9-a0d8-e55dc598e1b5 req-8ecf95f6-f27c-4e71-ab03-50d5f0e3da41 service nova] Releasing lock "refresh_cache-e9b2e5d1-e75e-4346-a0af-c0dafac0692e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.174068] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Acquired lock "refresh_cache-e9b2e5d1-e75e-4346-a0af-c0dafac0692e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.174262] env[61243]: DEBUG nova.network.neutron [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 803.578071] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e0626b-cead-4689-b596-95307b5e0ea5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.587095] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e68222a-17d0-4b57-b91e-7021492d15b7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.616819] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Releasing lock "refresh_cache-5314f922-6603-40dd-8760-9c1b77d6bd0a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.617064] env[61243]: DEBUG nova.compute.manager [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 803.617266] env[61243]: DEBUG nova.compute.manager [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 803.617438] env[61243]: DEBUG nova.network.neutron [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 803.620378] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b40ecf-03a1-46f7-bbeb-a3b1fab43884 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.630677] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b46fc6a-a71c-4040-b4ca-0e76d40e3b82 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.642670] env[61243]: DEBUG nova.compute.provider_tree [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.645179] env[61243]: DEBUG nova.network.neutron [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 803.695822] env[61243]: DEBUG nova.network.neutron [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 803.778020] env[61243]: DEBUG nova.network.neutron [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.072608] env[61243]: DEBUG nova.compute.manager [req-803c1b2a-44eb-4466-8879-2e526070bf0f req-58c5471d-e69f-4721-ab56-42b399516f4a service nova] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Received event network-vif-deleted-cff38ea8-81bd-4bd5-b704-b61286d84199 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.152918] env[61243]: DEBUG nova.scheduler.client.report [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.153590] env[61243]: DEBUG nova.network.neutron [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.283077] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Releasing lock "refresh_cache-e9b2e5d1-e75e-4346-a0af-c0dafac0692e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.283077] env[61243]: DEBUG nova.compute.manager [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 804.283077] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 804.283077] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dd670da0-bc83-4194-b778-4fdb2b3ec3d6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.291453] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e92e742-2a2f-4faf-bca3-96a47ee63f05 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.316738] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e9b2e5d1-e75e-4346-a0af-c0dafac0692e could not be found. [ 804.317341] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 804.317748] env[61243]: INFO nova.compute.manager [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 804.318529] env[61243]: DEBUG oslo.service.loopingcall [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.319107] env[61243]: DEBUG nova.compute.manager [-] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 804.319335] env[61243]: DEBUG nova.network.neutron [-] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 804.337284] env[61243]: DEBUG nova.network.neutron [-] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.662018] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.662018] env[61243]: DEBUG nova.compute.manager [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 804.663106] env[61243]: INFO nova.compute.manager [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] [instance: 5314f922-6603-40dd-8760-9c1b77d6bd0a] Took 1.05 seconds to deallocate network for instance. [ 804.666017] env[61243]: DEBUG oslo_concurrency.lockutils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.698s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.667571] env[61243]: INFO nova.compute.claims [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 804.841793] env[61243]: DEBUG nova.network.neutron [-] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.167717] env[61243]: DEBUG nova.compute.utils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 805.169148] env[61243]: DEBUG nova.compute.manager [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 805.169324] env[61243]: DEBUG nova.network.neutron [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 805.242060] env[61243]: DEBUG nova.policy [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4dc60c433f984e70917f48bab6c02b1c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '399030c8863346ea97f6da669ff3868b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 805.343442] env[61243]: INFO nova.compute.manager [-] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Took 1.02 seconds to deallocate network for instance. [ 805.346106] env[61243]: DEBUG nova.compute.claims [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Aborting claim: {{(pid=61243) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 805.346299] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.582354] env[61243]: DEBUG nova.network.neutron [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Successfully created port: f2248ec9-0230-4aa8-a016-a30ddf4f45f4 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 805.673634] env[61243]: DEBUG nova.compute.manager [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 805.712689] env[61243]: INFO nova.scheduler.client.report [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Deleted allocations for instance 5314f922-6603-40dd-8760-9c1b77d6bd0a [ 805.973919] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d601fb58-8e57-41d1-b0f9-be0a35cde134 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.981832] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-321cffc2-2b58-4280-bc41-9d2794cd1575 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.011862] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af04aec-fe25-4bd6-9f61-7b588c982085 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.019219] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32fcd571-9b92-45fd-a619-157155e996fa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.032229] env[61243]: DEBUG nova.compute.provider_tree [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.224723] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b5a3c59d-c362-441b-9b49-0143712021b8 tempest-ServerAddressesTestJSON-1817496302 tempest-ServerAddressesTestJSON-1817496302-project-member] Lock "5314f922-6603-40dd-8760-9c1b77d6bd0a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.787s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.536726] env[61243]: DEBUG nova.scheduler.client.report [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.689948] env[61243]: DEBUG nova.compute.manager [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 806.718516] env[61243]: DEBUG nova.virt.hardware [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 806.718766] env[61243]: DEBUG nova.virt.hardware [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 806.718922] env[61243]: DEBUG nova.virt.hardware [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.719118] env[61243]: DEBUG nova.virt.hardware [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 806.719269] env[61243]: DEBUG nova.virt.hardware [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.719416] env[61243]: DEBUG nova.virt.hardware [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 806.719620] env[61243]: DEBUG nova.virt.hardware [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 806.719781] env[61243]: DEBUG nova.virt.hardware [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 806.719941] env[61243]: DEBUG nova.virt.hardware [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 806.720167] env[61243]: DEBUG nova.virt.hardware [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 806.720357] env[61243]: DEBUG nova.virt.hardware [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 806.723927] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c324337-617f-411d-95bc-677ca24641fe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.726823] env[61243]: DEBUG nova.compute.manager [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 806.732636] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50767817-4830-46a3-8be2-fcc0fdadb5d2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.045916] env[61243]: DEBUG oslo_concurrency.lockutils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.046509] env[61243]: DEBUG nova.compute.manager [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 807.049542] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.190s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.050931] env[61243]: INFO nova.compute.claims [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 807.144463] env[61243]: DEBUG nova.compute.manager [req-ffe07ad1-e358-4f34-93f0-4495c54b0b83 req-c9a0c91c-4c55-49bf-9ae7-eee6f7907602 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Received event network-vif-plugged-f2248ec9-0230-4aa8-a016-a30ddf4f45f4 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 807.144765] env[61243]: DEBUG oslo_concurrency.lockutils [req-ffe07ad1-e358-4f34-93f0-4495c54b0b83 req-c9a0c91c-4c55-49bf-9ae7-eee6f7907602 service nova] Acquiring lock "86179fb0-99df-4b10-a815-c19168e9521e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.144988] env[61243]: DEBUG oslo_concurrency.lockutils [req-ffe07ad1-e358-4f34-93f0-4495c54b0b83 req-c9a0c91c-4c55-49bf-9ae7-eee6f7907602 service nova] Lock "86179fb0-99df-4b10-a815-c19168e9521e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.145327] env[61243]: DEBUG oslo_concurrency.lockutils [req-ffe07ad1-e358-4f34-93f0-4495c54b0b83 req-c9a0c91c-4c55-49bf-9ae7-eee6f7907602 service nova] Lock "86179fb0-99df-4b10-a815-c19168e9521e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.145551] env[61243]: DEBUG nova.compute.manager [req-ffe07ad1-e358-4f34-93f0-4495c54b0b83 req-c9a0c91c-4c55-49bf-9ae7-eee6f7907602 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] No waiting events found dispatching network-vif-plugged-f2248ec9-0230-4aa8-a016-a30ddf4f45f4 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 807.145764] env[61243]: WARNING nova.compute.manager [req-ffe07ad1-e358-4f34-93f0-4495c54b0b83 req-c9a0c91c-4c55-49bf-9ae7-eee6f7907602 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Received unexpected event network-vif-plugged-f2248ec9-0230-4aa8-a016-a30ddf4f45f4 for instance with vm_state building and task_state spawning. [ 807.235998] env[61243]: DEBUG nova.network.neutron [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Successfully updated port: f2248ec9-0230-4aa8-a016-a30ddf4f45f4 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 807.258644] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.564742] env[61243]: DEBUG nova.compute.utils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 807.571805] env[61243]: DEBUG nova.compute.manager [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 807.572110] env[61243]: DEBUG nova.network.neutron [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 807.619560] env[61243]: DEBUG nova.policy [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a96a8e8565f84ba1b2c2406e6a0f224b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8549626a07e941e1ab03452355a0f116', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 807.742607] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.742879] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquired lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.742930] env[61243]: DEBUG nova.network.neutron [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 807.918919] env[61243]: DEBUG nova.network.neutron [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Successfully created port: af7195a1-9a59-4b3a-b62d-7e907c94ce7b {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 808.071951] env[61243]: DEBUG nova.compute.manager [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 808.284175] env[61243]: DEBUG nova.network.neutron [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.416486] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-272158aa-757e-466c-a3a2-af5d9dd34129 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.424681] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c26542-8218-4e98-8ba5-116ccbe78d02 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.461031] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9072dfa8-263b-4937-8793-22465258e7c1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.466796] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a315ec-610a-4c1a-abed-252f83e31948 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.486746] env[61243]: DEBUG nova.compute.provider_tree [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.494051] env[61243]: DEBUG nova.network.neutron [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Updating instance_info_cache with network_info: [{"id": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "address": "fa:16:3e:b0:f1:66", "network": {"id": "49be247d-97e7-4085-8272-4ccf41767d10", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1335943921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "399030c8863346ea97f6da669ff3868b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2248ec9-02", "ovs_interfaceid": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.991023] env[61243]: DEBUG nova.scheduler.client.report [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.996429] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Releasing lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.996768] env[61243]: DEBUG nova.compute.manager [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Instance network_info: |[{"id": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "address": "fa:16:3e:b0:f1:66", "network": {"id": "49be247d-97e7-4085-8272-4ccf41767d10", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1335943921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "399030c8863346ea97f6da669ff3868b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2248ec9-02", "ovs_interfaceid": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 808.997341] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:f1:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e547d234-640c-449b-8279-0b16f75d6627', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f2248ec9-0230-4aa8-a016-a30ddf4f45f4', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 809.005618] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Creating folder: Project (399030c8863346ea97f6da669ff3868b). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.006658] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3761f331-8bb2-48f3-8f31-102c276f4f37 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.019273] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Created folder: Project (399030c8863346ea97f6da669ff3868b) in parent group-v285636. [ 809.019523] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Creating folder: Instances. Parent ref: group-v285656. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.019706] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f5e36181-2123-4b42-a553-cbc71d9aa03b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.030131] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Created folder: Instances in parent group-v285656. [ 809.030375] env[61243]: DEBUG oslo.service.loopingcall [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.030946] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 809.030946] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fec2021a-17e3-488c-9e89-16f3c70a7745 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.055981] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 809.055981] env[61243]: value = "task-1338762" [ 809.055981] env[61243]: _type = "Task" [ 809.055981] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.063781] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338762, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.083590] env[61243]: DEBUG nova.compute.manager [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 809.115203] env[61243]: DEBUG nova.virt.hardware [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 809.117104] env[61243]: DEBUG nova.virt.hardware [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 809.117104] env[61243]: DEBUG nova.virt.hardware [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 809.117104] env[61243]: DEBUG nova.virt.hardware [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 809.117104] env[61243]: DEBUG nova.virt.hardware [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 809.117104] env[61243]: DEBUG nova.virt.hardware [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 809.117532] env[61243]: DEBUG nova.virt.hardware [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 809.118118] env[61243]: DEBUG nova.virt.hardware [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 809.118647] env[61243]: DEBUG nova.virt.hardware [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 809.118809] env[61243]: DEBUG nova.virt.hardware [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 809.119075] env[61243]: DEBUG nova.virt.hardware [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 809.120588] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f58d69-3d27-49c7-8bc4-ad819c5e4027 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.129803] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b4efe3-afe1-4a12-8883-9119e7e9911f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.177409] env[61243]: DEBUG nova.compute.manager [req-1615c327-e40d-4708-97cb-e4e8778fecc8 req-3236dbce-be21-4543-839c-cd59de680c9b service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Received event network-changed-f2248ec9-0230-4aa8-a016-a30ddf4f45f4 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.177685] env[61243]: DEBUG nova.compute.manager [req-1615c327-e40d-4708-97cb-e4e8778fecc8 req-3236dbce-be21-4543-839c-cd59de680c9b service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Refreshing instance network info cache due to event network-changed-f2248ec9-0230-4aa8-a016-a30ddf4f45f4. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 809.177994] env[61243]: DEBUG oslo_concurrency.lockutils [req-1615c327-e40d-4708-97cb-e4e8778fecc8 req-3236dbce-be21-4543-839c-cd59de680c9b service nova] Acquiring lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.178210] env[61243]: DEBUG oslo_concurrency.lockutils [req-1615c327-e40d-4708-97cb-e4e8778fecc8 req-3236dbce-be21-4543-839c-cd59de680c9b service nova] Acquired lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.178467] env[61243]: DEBUG nova.network.neutron [req-1615c327-e40d-4708-97cb-e4e8778fecc8 req-3236dbce-be21-4543-839c-cd59de680c9b service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Refreshing network info cache for port f2248ec9-0230-4aa8-a016-a30ddf4f45f4 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 809.496869] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.497501] env[61243]: DEBUG nova.compute.manager [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 809.500597] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.602s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.502503] env[61243]: INFO nova.compute.claims [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.567046] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338762, 'name': CreateVM_Task, 'duration_secs': 0.320689} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.567309] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 809.576411] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.576634] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.578017] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 809.578017] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ceecceb5-9612-4404-bb78-86a88da8b68a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.585714] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 809.585714] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a4c978-4c06-7fc4-51ec-a91117dcf3ae" [ 809.585714] env[61243]: _type = "Task" [ 809.585714] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.594569] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a4c978-4c06-7fc4-51ec-a91117dcf3ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.651080] env[61243]: DEBUG nova.network.neutron [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Successfully updated port: af7195a1-9a59-4b3a-b62d-7e907c94ce7b {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 809.892082] env[61243]: DEBUG nova.network.neutron [req-1615c327-e40d-4708-97cb-e4e8778fecc8 req-3236dbce-be21-4543-839c-cd59de680c9b service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Updated VIF entry in instance network info cache for port f2248ec9-0230-4aa8-a016-a30ddf4f45f4. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 809.892486] env[61243]: DEBUG nova.network.neutron [req-1615c327-e40d-4708-97cb-e4e8778fecc8 req-3236dbce-be21-4543-839c-cd59de680c9b service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Updating instance_info_cache with network_info: [{"id": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "address": "fa:16:3e:b0:f1:66", "network": {"id": "49be247d-97e7-4085-8272-4ccf41767d10", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1335943921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "399030c8863346ea97f6da669ff3868b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2248ec9-02", "ovs_interfaceid": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.010022] env[61243]: DEBUG nova.compute.utils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 810.012715] env[61243]: DEBUG nova.compute.manager [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 810.013048] env[61243]: DEBUG nova.network.neutron [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 810.070509] env[61243]: DEBUG nova.policy [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cc77f359ba44dc0bc8e053d5a4b63ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3248cca8a8aa4cad84b8bd5dcf7ed0c3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 810.099747] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a4c978-4c06-7fc4-51ec-a91117dcf3ae, 'name': SearchDatastore_Task, 'duration_secs': 0.013204} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.100804] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.100804] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 810.100804] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.100804] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.101042] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 810.101491] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eb7ffed0-1a86-431b-badf-89a66da3d9b0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.112822] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 810.112822] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 810.112822] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa5c7cdd-8929-44d7-8b87-34155888bcf0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.115990] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 810.115990] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522bad05-87bb-e7e0-616b-c6a6cc814d10" [ 810.115990] env[61243]: _type = "Task" [ 810.115990] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.126135] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522bad05-87bb-e7e0-616b-c6a6cc814d10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.158072] env[61243]: DEBUG oslo_concurrency.lockutils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.158072] env[61243]: DEBUG oslo_concurrency.lockutils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquired lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.158072] env[61243]: DEBUG nova.network.neutron [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 810.344259] env[61243]: DEBUG nova.network.neutron [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Successfully created port: 471f4f83-e6c9-47ce-bfde-c78219c698a6 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.395612] env[61243]: DEBUG oslo_concurrency.lockutils [req-1615c327-e40d-4708-97cb-e4e8778fecc8 req-3236dbce-be21-4543-839c-cd59de680c9b service nova] Releasing lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.517029] env[61243]: DEBUG nova.compute.manager [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 810.632943] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522bad05-87bb-e7e0-616b-c6a6cc814d10, 'name': SearchDatastore_Task, 'duration_secs': 0.010676} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.634950] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83d995a0-d0d0-442c-aaa7-3421e341f518 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.643363] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 810.643363] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52423e7c-9dcf-2571-c253-38d8886427e9" [ 810.643363] env[61243]: _type = "Task" [ 810.643363] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.649985] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52423e7c-9dcf-2571-c253-38d8886427e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.700912] env[61243]: DEBUG nova.network.neutron [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 810.807341] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0719ce73-ec90-409a-a70a-c1442645bcfe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.815576] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e697cb-c169-4c0b-a7dc-bbd40d4aa5cb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.848352] env[61243]: DEBUG nova.network.neutron [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Updating instance_info_cache with network_info: [{"id": "af7195a1-9a59-4b3a-b62d-7e907c94ce7b", "address": "fa:16:3e:ee:b7:e1", "network": {"id": "c6d4943d-4a8d-41b4-8023-c4f1549d312d", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1045259606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8549626a07e941e1ab03452355a0f116", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf7195a1-9a", "ovs_interfaceid": "af7195a1-9a59-4b3a-b62d-7e907c94ce7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.850188] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c191fc-4404-44f8-b7ab-822fa544ce8f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.857987] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31dd0836-5295-4b16-b0ac-0d998e072665 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.877796] env[61243]: DEBUG nova.compute.provider_tree [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.152107] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52423e7c-9dcf-2571-c253-38d8886427e9, 'name': SearchDatastore_Task, 'duration_secs': 0.009579} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.152429] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.152667] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 86179fb0-99df-4b10-a815-c19168e9521e/86179fb0-99df-4b10-a815-c19168e9521e.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 811.152939] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-efc6c9ee-727d-4268-ad78-5d18382ef761 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.159591] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 811.159591] env[61243]: value = "task-1338763" [ 811.159591] env[61243]: _type = "Task" [ 811.159591] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.172164] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338763, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.235114] env[61243]: DEBUG nova.compute.manager [req-9ec853ef-b5df-46fb-8c04-660dba5e8f94 req-6dc6c3d0-83e1-4883-a43a-eef5f5d2626a service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Received event network-vif-plugged-af7195a1-9a59-4b3a-b62d-7e907c94ce7b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.235421] env[61243]: DEBUG oslo_concurrency.lockutils [req-9ec853ef-b5df-46fb-8c04-660dba5e8f94 req-6dc6c3d0-83e1-4883-a43a-eef5f5d2626a service nova] Acquiring lock "07ee984d-476e-484d-ba80-0ec2e411faa9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.235952] env[61243]: DEBUG oslo_concurrency.lockutils [req-9ec853ef-b5df-46fb-8c04-660dba5e8f94 req-6dc6c3d0-83e1-4883-a43a-eef5f5d2626a service nova] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.236018] env[61243]: DEBUG oslo_concurrency.lockutils [req-9ec853ef-b5df-46fb-8c04-660dba5e8f94 req-6dc6c3d0-83e1-4883-a43a-eef5f5d2626a service nova] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.236345] env[61243]: DEBUG nova.compute.manager [req-9ec853ef-b5df-46fb-8c04-660dba5e8f94 req-6dc6c3d0-83e1-4883-a43a-eef5f5d2626a service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] No waiting events found dispatching network-vif-plugged-af7195a1-9a59-4b3a-b62d-7e907c94ce7b {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 811.236567] env[61243]: WARNING nova.compute.manager [req-9ec853ef-b5df-46fb-8c04-660dba5e8f94 req-6dc6c3d0-83e1-4883-a43a-eef5f5d2626a service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Received unexpected event network-vif-plugged-af7195a1-9a59-4b3a-b62d-7e907c94ce7b for instance with vm_state building and task_state spawning. [ 811.237141] env[61243]: DEBUG nova.compute.manager [req-9ec853ef-b5df-46fb-8c04-660dba5e8f94 req-6dc6c3d0-83e1-4883-a43a-eef5f5d2626a service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Received event network-changed-af7195a1-9a59-4b3a-b62d-7e907c94ce7b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.237141] env[61243]: DEBUG nova.compute.manager [req-9ec853ef-b5df-46fb-8c04-660dba5e8f94 req-6dc6c3d0-83e1-4883-a43a-eef5f5d2626a service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Refreshing instance network info cache due to event network-changed-af7195a1-9a59-4b3a-b62d-7e907c94ce7b. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 811.237220] env[61243]: DEBUG oslo_concurrency.lockutils [req-9ec853ef-b5df-46fb-8c04-660dba5e8f94 req-6dc6c3d0-83e1-4883-a43a-eef5f5d2626a service nova] Acquiring lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.354345] env[61243]: DEBUG oslo_concurrency.lockutils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Releasing lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.354867] env[61243]: DEBUG nova.compute.manager [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Instance network_info: |[{"id": "af7195a1-9a59-4b3a-b62d-7e907c94ce7b", "address": "fa:16:3e:ee:b7:e1", "network": {"id": "c6d4943d-4a8d-41b4-8023-c4f1549d312d", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1045259606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8549626a07e941e1ab03452355a0f116", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf7195a1-9a", "ovs_interfaceid": "af7195a1-9a59-4b3a-b62d-7e907c94ce7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 811.355207] env[61243]: DEBUG oslo_concurrency.lockutils [req-9ec853ef-b5df-46fb-8c04-660dba5e8f94 req-6dc6c3d0-83e1-4883-a43a-eef5f5d2626a service nova] Acquired lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.355400] env[61243]: DEBUG nova.network.neutron [req-9ec853ef-b5df-46fb-8c04-660dba5e8f94 req-6dc6c3d0-83e1-4883-a43a-eef5f5d2626a service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Refreshing network info cache for port af7195a1-9a59-4b3a-b62d-7e907c94ce7b {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 811.356742] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:b7:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24ec44b7-0acf-4ff9-8bb3-4641b74af7a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'af7195a1-9a59-4b3a-b62d-7e907c94ce7b', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 811.368417] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Creating folder: Project (8549626a07e941e1ab03452355a0f116). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.368417] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5940e51b-a91d-4cdf-a779-caec3b77ec5e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.379332] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Created folder: Project (8549626a07e941e1ab03452355a0f116) in parent group-v285636. [ 811.379644] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Creating folder: Instances. Parent ref: group-v285659. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.379863] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0e26dea9-0c62-4526-926a-98dc5e0d0268 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.383917] env[61243]: DEBUG nova.scheduler.client.report [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.395896] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Created folder: Instances in parent group-v285659. [ 811.396205] env[61243]: DEBUG oslo.service.loopingcall [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.396414] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 811.396636] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-77a50d11-7c14-4be3-839e-55cd8a772827 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.419599] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 811.419599] env[61243]: value = "task-1338766" [ 811.419599] env[61243]: _type = "Task" [ 811.419599] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.430631] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338766, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.528119] env[61243]: DEBUG nova.compute.manager [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 811.562550] env[61243]: DEBUG nova.virt.hardware [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.562855] env[61243]: DEBUG nova.virt.hardware [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.563157] env[61243]: DEBUG nova.virt.hardware [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.563382] env[61243]: DEBUG nova.virt.hardware [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.563612] env[61243]: DEBUG nova.virt.hardware [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.563710] env[61243]: DEBUG nova.virt.hardware [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.563892] env[61243]: DEBUG nova.virt.hardware [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.564067] env[61243]: DEBUG nova.virt.hardware [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.564239] env[61243]: DEBUG nova.virt.hardware [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.564406] env[61243]: DEBUG nova.virt.hardware [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.564581] env[61243]: DEBUG nova.virt.hardware [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.565669] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc128c1a-a4be-455e-bc83-2473887ee476 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.574909] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c75f45-65d5-4a94-b28e-297adc48a8a6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.669694] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338763, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.855212] env[61243]: DEBUG nova.network.neutron [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Successfully updated port: 471f4f83-e6c9-47ce-bfde-c78219c698a6 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 811.889377] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.889885] env[61243]: DEBUG nova.compute.manager [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 811.893645] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.065s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.896380] env[61243]: INFO nova.compute.claims [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.931443] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338766, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.115675] env[61243]: DEBUG nova.network.neutron [req-9ec853ef-b5df-46fb-8c04-660dba5e8f94 req-6dc6c3d0-83e1-4883-a43a-eef5f5d2626a service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Updated VIF entry in instance network info cache for port af7195a1-9a59-4b3a-b62d-7e907c94ce7b. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 812.117022] env[61243]: DEBUG nova.network.neutron [req-9ec853ef-b5df-46fb-8c04-660dba5e8f94 req-6dc6c3d0-83e1-4883-a43a-eef5f5d2626a service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Updating instance_info_cache with network_info: [{"id": "af7195a1-9a59-4b3a-b62d-7e907c94ce7b", "address": "fa:16:3e:ee:b7:e1", "network": {"id": "c6d4943d-4a8d-41b4-8023-c4f1549d312d", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1045259606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8549626a07e941e1ab03452355a0f116", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf7195a1-9a", "ovs_interfaceid": "af7195a1-9a59-4b3a-b62d-7e907c94ce7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.170849] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338763, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510853} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.171146] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 86179fb0-99df-4b10-a815-c19168e9521e/86179fb0-99df-4b10-a815-c19168e9521e.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 812.171331] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 812.171581] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e8af3fe1-1f89-40b8-b030-370315a5b994 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.177800] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 812.177800] env[61243]: value = "task-1338767" [ 812.177800] env[61243]: _type = "Task" [ 812.177800] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.186118] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338767, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.359537] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "refresh_cache-36e55334-8628-4dd7-a845-f4ae3d8e7ff9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.359537] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired lock "refresh_cache-36e55334-8628-4dd7-a845-f4ae3d8e7ff9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.359537] env[61243]: DEBUG nova.network.neutron [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 812.402318] env[61243]: DEBUG nova.compute.utils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.403690] env[61243]: DEBUG nova.compute.manager [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 812.403858] env[61243]: DEBUG nova.network.neutron [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 812.431714] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338766, 'name': CreateVM_Task, 'duration_secs': 0.527812} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.431823] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 812.432516] env[61243]: DEBUG oslo_concurrency.lockutils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.432668] env[61243]: DEBUG oslo_concurrency.lockutils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.433555] env[61243]: DEBUG oslo_concurrency.lockutils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 812.433555] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3b28996-f6dd-44a0-8251-5b90c8581a15 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.438423] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 812.438423] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e5fcf9-ec11-c4f4-b522-ab018a6bc364" [ 812.438423] env[61243]: _type = "Task" [ 812.438423] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.446467] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e5fcf9-ec11-c4f4-b522-ab018a6bc364, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.448169] env[61243]: DEBUG nova.policy [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b32f4788825f4a8e82e028cc10347746', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '683c027e4adc4cc4a6b8747a8c8c0de9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 812.618732] env[61243]: DEBUG oslo_concurrency.lockutils [req-9ec853ef-b5df-46fb-8c04-660dba5e8f94 req-6dc6c3d0-83e1-4883-a43a-eef5f5d2626a service nova] Releasing lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.691244] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338767, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.134745} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.691625] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 812.692435] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25fdc3c7-7988-434c-bf53-3d6727bfd81e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.714091] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 86179fb0-99df-4b10-a815-c19168e9521e/86179fb0-99df-4b10-a815-c19168e9521e.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.714394] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a15e043-392a-4238-959e-bd8ffc45c9ee {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.736338] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 812.736338] env[61243]: value = "task-1338768" [ 812.736338] env[61243]: _type = "Task" [ 812.736338] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.744753] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338768, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.745698] env[61243]: DEBUG nova.network.neutron [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Successfully created port: 4cd8c409-8daa-4f53-ac5f-d2ac25de247b {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 812.902450] env[61243]: DEBUG nova.network.neutron [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 812.910122] env[61243]: DEBUG nova.compute.manager [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 812.950296] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e5fcf9-ec11-c4f4-b522-ab018a6bc364, 'name': SearchDatastore_Task, 'duration_secs': 0.008562} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.952788] env[61243]: DEBUG oslo_concurrency.lockutils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.953322] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 812.953426] env[61243]: DEBUG oslo_concurrency.lockutils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.953514] env[61243]: DEBUG oslo_concurrency.lockutils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.953795] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.954376] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b2e73ec-1937-4cb5-946e-81512fc725e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.965297] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.965523] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 812.969885] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4be30c0d-d16b-433a-bb00-c057d033760e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.985294] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 812.985294] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5202f1fa-1ee6-3856-4c59-a86fcabf6655" [ 812.985294] env[61243]: _type = "Task" [ 812.985294] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.000729] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5202f1fa-1ee6-3856-4c59-a86fcabf6655, 'name': SearchDatastore_Task, 'duration_secs': 0.00919} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.004609] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6ff0365-18f5-4337-a465-00701fe769b8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.010127] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 813.010127] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]526b0d97-89ef-931d-14d5-17bc1ca02c2e" [ 813.010127] env[61243]: _type = "Task" [ 813.010127] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.022855] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]526b0d97-89ef-931d-14d5-17bc1ca02c2e, 'name': SearchDatastore_Task} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.026456] env[61243]: DEBUG oslo_concurrency.lockutils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.026456] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 07ee984d-476e-484d-ba80-0ec2e411faa9/07ee984d-476e-484d-ba80-0ec2e411faa9.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 813.028066] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e4d58709-5e0f-4da6-a401-ac4cd7b7a5d8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.035311] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 813.035311] env[61243]: value = "task-1338769" [ 813.035311] env[61243]: _type = "Task" [ 813.035311] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.049138] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338769, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.116577] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Acquiring lock "f73a4d39-5478-4135-9be6-e59f3e29788d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.117055] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Lock "f73a4d39-5478-4135-9be6-e59f3e29788d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.140613] env[61243]: DEBUG nova.network.neutron [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Updating instance_info_cache with network_info: [{"id": "471f4f83-e6c9-47ce-bfde-c78219c698a6", "address": "fa:16:3e:51:76:d2", "network": {"id": "2737f3f1-a8da-4b5f-a33d-c768f14207e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1004846100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3248cca8a8aa4cad84b8bd5dcf7ed0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap471f4f83-e6", "ovs_interfaceid": "471f4f83-e6c9-47ce-bfde-c78219c698a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.249628] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338768, 'name': ReconfigVM_Task, 'duration_secs': 0.309603} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.250251] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 86179fb0-99df-4b10-a815-c19168e9521e/86179fb0-99df-4b10-a815-c19168e9521e.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.250576] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ea2ffbd4-5acd-4d4e-963f-e59590100920 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.254626] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a662af-8a44-4849-9ebf-58ea17374522 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.258152] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 813.258152] env[61243]: value = "task-1338770" [ 813.258152] env[61243]: _type = "Task" [ 813.258152] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.265760] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2923c276-9042-42a7-926a-8bb918c285ca {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.269905] env[61243]: DEBUG nova.compute.manager [req-34082854-14a8-4ef0-8b08-9de349992ba6 req-2eed6929-4a3b-4f3f-b0e1-a88553959856 service nova] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Received event network-vif-plugged-471f4f83-e6c9-47ce-bfde-c78219c698a6 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 813.270150] env[61243]: DEBUG oslo_concurrency.lockutils [req-34082854-14a8-4ef0-8b08-9de349992ba6 req-2eed6929-4a3b-4f3f-b0e1-a88553959856 service nova] Acquiring lock "36e55334-8628-4dd7-a845-f4ae3d8e7ff9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.270461] env[61243]: DEBUG oslo_concurrency.lockutils [req-34082854-14a8-4ef0-8b08-9de349992ba6 req-2eed6929-4a3b-4f3f-b0e1-a88553959856 service nova] Lock "36e55334-8628-4dd7-a845-f4ae3d8e7ff9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.270634] env[61243]: DEBUG oslo_concurrency.lockutils [req-34082854-14a8-4ef0-8b08-9de349992ba6 req-2eed6929-4a3b-4f3f-b0e1-a88553959856 service nova] Lock "36e55334-8628-4dd7-a845-f4ae3d8e7ff9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.271073] env[61243]: DEBUG nova.compute.manager [req-34082854-14a8-4ef0-8b08-9de349992ba6 req-2eed6929-4a3b-4f3f-b0e1-a88553959856 service nova] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] No waiting events found dispatching network-vif-plugged-471f4f83-e6c9-47ce-bfde-c78219c698a6 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 813.271477] env[61243]: WARNING nova.compute.manager [req-34082854-14a8-4ef0-8b08-9de349992ba6 req-2eed6929-4a3b-4f3f-b0e1-a88553959856 service nova] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Received unexpected event network-vif-plugged-471f4f83-e6c9-47ce-bfde-c78219c698a6 for instance with vm_state building and task_state spawning. [ 813.271696] env[61243]: DEBUG nova.compute.manager [req-34082854-14a8-4ef0-8b08-9de349992ba6 req-2eed6929-4a3b-4f3f-b0e1-a88553959856 service nova] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Received event network-changed-471f4f83-e6c9-47ce-bfde-c78219c698a6 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 813.271868] env[61243]: DEBUG nova.compute.manager [req-34082854-14a8-4ef0-8b08-9de349992ba6 req-2eed6929-4a3b-4f3f-b0e1-a88553959856 service nova] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Refreshing instance network info cache due to event network-changed-471f4f83-e6c9-47ce-bfde-c78219c698a6. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 813.272060] env[61243]: DEBUG oslo_concurrency.lockutils [req-34082854-14a8-4ef0-8b08-9de349992ba6 req-2eed6929-4a3b-4f3f-b0e1-a88553959856 service nova] Acquiring lock "refresh_cache-36e55334-8628-4dd7-a845-f4ae3d8e7ff9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.276518] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338770, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.311847] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6afc7472-2312-43e9-9d17-15c6100ccfbf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.328305] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847ae04d-07a4-4164-8fe3-71b1ad150f24 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.340321] env[61243]: DEBUG nova.compute.provider_tree [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.546526] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338769, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.646253] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Releasing lock "refresh_cache-36e55334-8628-4dd7-a845-f4ae3d8e7ff9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.646253] env[61243]: DEBUG nova.compute.manager [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Instance network_info: |[{"id": "471f4f83-e6c9-47ce-bfde-c78219c698a6", "address": "fa:16:3e:51:76:d2", "network": {"id": "2737f3f1-a8da-4b5f-a33d-c768f14207e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1004846100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3248cca8a8aa4cad84b8bd5dcf7ed0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap471f4f83-e6", "ovs_interfaceid": "471f4f83-e6c9-47ce-bfde-c78219c698a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 813.646741] env[61243]: DEBUG oslo_concurrency.lockutils [req-34082854-14a8-4ef0-8b08-9de349992ba6 req-2eed6929-4a3b-4f3f-b0e1-a88553959856 service nova] Acquired lock "refresh_cache-36e55334-8628-4dd7-a845-f4ae3d8e7ff9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.646741] env[61243]: DEBUG nova.network.neutron [req-34082854-14a8-4ef0-8b08-9de349992ba6 req-2eed6929-4a3b-4f3f-b0e1-a88553959856 service nova] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Refreshing network info cache for port 471f4f83-e6c9-47ce-bfde-c78219c698a6 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 813.648503] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:76:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '26472e27-9835-4f87-ab7f-ca24dfee4e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '471f4f83-e6c9-47ce-bfde-c78219c698a6', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 813.655594] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Creating folder: Project (3248cca8a8aa4cad84b8bd5dcf7ed0c3). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 813.659299] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e13e32ab-d313-4f28-b0a5-e84f04c90528 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.671810] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Created folder: Project (3248cca8a8aa4cad84b8bd5dcf7ed0c3) in parent group-v285636. [ 813.671810] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Creating folder: Instances. Parent ref: group-v285662. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 813.671810] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6b4cd90a-cc49-492d-8f03-cf534e052f7d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.679645] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Created folder: Instances in parent group-v285662. [ 813.679722] env[61243]: DEBUG oslo.service.loopingcall [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.680698] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 813.680698] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1f173926-68dd-4d0d-9681-a8047e5c013b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.699933] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 813.699933] env[61243]: value = "task-1338773" [ 813.699933] env[61243]: _type = "Task" [ 813.699933] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.711840] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338773, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.767596] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338770, 'name': Rename_Task, 'duration_secs': 0.232667} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.767878] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 813.768163] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eaa0d9c8-5de0-47c1-a517-04f65f2d50a3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.773865] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 813.773865] env[61243]: value = "task-1338774" [ 813.773865] env[61243]: _type = "Task" [ 813.773865] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.784449] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338774, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.845036] env[61243]: DEBUG nova.scheduler.client.report [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.885639] env[61243]: DEBUG nova.network.neutron [req-34082854-14a8-4ef0-8b08-9de349992ba6 req-2eed6929-4a3b-4f3f-b0e1-a88553959856 service nova] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Updated VIF entry in instance network info cache for port 471f4f83-e6c9-47ce-bfde-c78219c698a6. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 813.886046] env[61243]: DEBUG nova.network.neutron [req-34082854-14a8-4ef0-8b08-9de349992ba6 req-2eed6929-4a3b-4f3f-b0e1-a88553959856 service nova] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Updating instance_info_cache with network_info: [{"id": "471f4f83-e6c9-47ce-bfde-c78219c698a6", "address": "fa:16:3e:51:76:d2", "network": {"id": "2737f3f1-a8da-4b5f-a33d-c768f14207e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1004846100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3248cca8a8aa4cad84b8bd5dcf7ed0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap471f4f83-e6", "ovs_interfaceid": "471f4f83-e6c9-47ce-bfde-c78219c698a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.924959] env[61243]: DEBUG nova.compute.manager [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 813.956918] env[61243]: DEBUG nova.virt.hardware [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 813.956918] env[61243]: DEBUG nova.virt.hardware [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 813.956918] env[61243]: DEBUG nova.virt.hardware [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 813.957261] env[61243]: DEBUG nova.virt.hardware [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 813.957261] env[61243]: DEBUG nova.virt.hardware [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 813.957261] env[61243]: DEBUG nova.virt.hardware [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 813.957261] env[61243]: DEBUG nova.virt.hardware [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 813.957261] env[61243]: DEBUG nova.virt.hardware [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 813.957409] env[61243]: DEBUG nova.virt.hardware [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 813.957409] env[61243]: DEBUG nova.virt.hardware [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 813.957409] env[61243]: DEBUG nova.virt.hardware [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 813.958179] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e260a82a-4886-4e57-b298-2e48a56d216f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.968188] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e090f8a3-84a8-4cee-b188-97e8f06d61f4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.046726] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338769, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542308} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.047132] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 07ee984d-476e-484d-ba80-0ec2e411faa9/07ee984d-476e-484d-ba80-0ec2e411faa9.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 814.047451] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 814.047768] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ff174420-1875-4b79-b6f4-17fb2bbbebda {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.054544] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 814.054544] env[61243]: value = "task-1338775" [ 814.054544] env[61243]: _type = "Task" [ 814.054544] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.062045] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338775, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.213772] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338773, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.283944] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338774, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.350813] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.458s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.351141] env[61243]: DEBUG nova.compute.manager [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 814.354327] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.982s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.389703] env[61243]: DEBUG oslo_concurrency.lockutils [req-34082854-14a8-4ef0-8b08-9de349992ba6 req-2eed6929-4a3b-4f3f-b0e1-a88553959856 service nova] Releasing lock "refresh_cache-36e55334-8628-4dd7-a845-f4ae3d8e7ff9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.390899] env[61243]: DEBUG nova.network.neutron [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Successfully updated port: 4cd8c409-8daa-4f53-ac5f-d2ac25de247b {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 814.569325] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338775, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.233732} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.569715] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 814.571179] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b768af-da17-4e49-9c34-690c94f1f5f5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.596034] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] 07ee984d-476e-484d-ba80-0ec2e411faa9/07ee984d-476e-484d-ba80-0ec2e411faa9.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 814.596303] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-850eb962-bd18-426a-a7c6-f5318c7b93bb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.619425] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 814.619425] env[61243]: value = "task-1338776" [ 814.619425] env[61243]: _type = "Task" [ 814.619425] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.630189] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338776, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.714945] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338773, 'name': CreateVM_Task, 'duration_secs': 0.530227} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.714945] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 814.715805] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.716065] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.716482] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 814.716818] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9080a039-6e9b-4b8d-bbe5-26fc50ed8da8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.722703] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 814.722703] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]520fa6e0-11f3-a0b8-2ce2-e375708f0674" [ 814.722703] env[61243]: _type = "Task" [ 814.722703] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.737386] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]520fa6e0-11f3-a0b8-2ce2-e375708f0674, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.784813] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338774, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.858567] env[61243]: DEBUG nova.compute.utils [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 814.861050] env[61243]: DEBUG nova.compute.manager [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Not allocating networking since 'none' was specified. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 814.893988] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.894146] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.894295] env[61243]: DEBUG nova.network.neutron [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.133715] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338776, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.137823] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a8b5937-1410-4cbb-8f56-9c864f4d0157 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.145535] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4afb0dd8-b475-442a-ab4c-2a3182aa88db {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.184510] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f204952a-4382-44e7-af95-cae21808a6be {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.193142] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f88f7722-666a-40e9-8b14-4ca7a26daf46 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.207934] env[61243]: DEBUG nova.compute.provider_tree [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.233043] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]520fa6e0-11f3-a0b8-2ce2-e375708f0674, 'name': SearchDatastore_Task, 'duration_secs': 0.012132} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.233370] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.233613] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 815.233918] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.234138] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.234345] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 815.234611] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-44ad12bf-25d5-4066-9aa9-d6bcdaa3c45b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.243690] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 815.243890] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 815.244639] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f0db5d6-ee2e-4571-b5d2-4294ab0af456 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.249790] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 815.249790] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f08ea1-54d6-7b42-fdd0-d310f4d8879f" [ 815.249790] env[61243]: _type = "Task" [ 815.249790] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.259345] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f08ea1-54d6-7b42-fdd0-d310f4d8879f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.285108] env[61243]: DEBUG oslo_vmware.api [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338774, 'name': PowerOnVM_Task, 'duration_secs': 1.019371} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.285415] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 815.285612] env[61243]: INFO nova.compute.manager [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Took 8.60 seconds to spawn the instance on the hypervisor. [ 815.285787] env[61243]: DEBUG nova.compute.manager [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 815.286620] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dfbc950-2cfa-4775-bdf3-e80c92793008 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.364812] env[61243]: DEBUG nova.compute.manager [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 815.387657] env[61243]: DEBUG nova.compute.manager [req-55811a87-7039-4ee8-a6d8-1eedee8a12f0 req-cba86aa3-45f6-46c8-81df-d2c68bb367ce service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Received event network-vif-plugged-4cd8c409-8daa-4f53-ac5f-d2ac25de247b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 815.387967] env[61243]: DEBUG oslo_concurrency.lockutils [req-55811a87-7039-4ee8-a6d8-1eedee8a12f0 req-cba86aa3-45f6-46c8-81df-d2c68bb367ce service nova] Acquiring lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.388247] env[61243]: DEBUG oslo_concurrency.lockutils [req-55811a87-7039-4ee8-a6d8-1eedee8a12f0 req-cba86aa3-45f6-46c8-81df-d2c68bb367ce service nova] Lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.388247] env[61243]: DEBUG oslo_concurrency.lockutils [req-55811a87-7039-4ee8-a6d8-1eedee8a12f0 req-cba86aa3-45f6-46c8-81df-d2c68bb367ce service nova] Lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.388388] env[61243]: DEBUG nova.compute.manager [req-55811a87-7039-4ee8-a6d8-1eedee8a12f0 req-cba86aa3-45f6-46c8-81df-d2c68bb367ce service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] No waiting events found dispatching network-vif-plugged-4cd8c409-8daa-4f53-ac5f-d2ac25de247b {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 815.388551] env[61243]: WARNING nova.compute.manager [req-55811a87-7039-4ee8-a6d8-1eedee8a12f0 req-cba86aa3-45f6-46c8-81df-d2c68bb367ce service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Received unexpected event network-vif-plugged-4cd8c409-8daa-4f53-ac5f-d2ac25de247b for instance with vm_state building and task_state spawning. [ 815.388714] env[61243]: DEBUG nova.compute.manager [req-55811a87-7039-4ee8-a6d8-1eedee8a12f0 req-cba86aa3-45f6-46c8-81df-d2c68bb367ce service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Received event network-changed-4cd8c409-8daa-4f53-ac5f-d2ac25de247b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 815.389054] env[61243]: DEBUG nova.compute.manager [req-55811a87-7039-4ee8-a6d8-1eedee8a12f0 req-cba86aa3-45f6-46c8-81df-d2c68bb367ce service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Refreshing instance network info cache due to event network-changed-4cd8c409-8daa-4f53-ac5f-d2ac25de247b. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 815.389054] env[61243]: DEBUG oslo_concurrency.lockutils [req-55811a87-7039-4ee8-a6d8-1eedee8a12f0 req-cba86aa3-45f6-46c8-81df-d2c68bb367ce service nova] Acquiring lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.428466] env[61243]: DEBUG nova.network.neutron [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.609825] env[61243]: DEBUG nova.network.neutron [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Updating instance_info_cache with network_info: [{"id": "4cd8c409-8daa-4f53-ac5f-d2ac25de247b", "address": "fa:16:3e:11:60:2d", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4cd8c409-8d", "ovs_interfaceid": "4cd8c409-8daa-4f53-ac5f-d2ac25de247b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.637232] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338776, 'name': ReconfigVM_Task, 'duration_secs': 0.966898} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.637677] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Reconfigured VM instance instance-00000038 to attach disk [datastore2] 07ee984d-476e-484d-ba80-0ec2e411faa9/07ee984d-476e-484d-ba80-0ec2e411faa9.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.638587] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9eaf4907-28d9-44c9-ab70-16ae1b29a930 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.646805] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 815.646805] env[61243]: value = "task-1338777" [ 815.646805] env[61243]: _type = "Task" [ 815.646805] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.659927] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338777, 'name': Rename_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.711589] env[61243]: DEBUG nova.scheduler.client.report [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.760206] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f08ea1-54d6-7b42-fdd0-d310f4d8879f, 'name': SearchDatastore_Task, 'duration_secs': 0.012024} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.761107] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-127693e6-4301-4644-877b-b56f3069e4f7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.766288] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 815.766288] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a8946c-79e6-767c-c18b-243e2bac727c" [ 815.766288] env[61243]: _type = "Task" [ 815.766288] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.774135] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a8946c-79e6-767c-c18b-243e2bac727c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.805742] env[61243]: INFO nova.compute.manager [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Took 37.83 seconds to build instance. [ 816.114372] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.114715] env[61243]: DEBUG nova.compute.manager [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Instance network_info: |[{"id": "4cd8c409-8daa-4f53-ac5f-d2ac25de247b", "address": "fa:16:3e:11:60:2d", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4cd8c409-8d", "ovs_interfaceid": "4cd8c409-8daa-4f53-ac5f-d2ac25de247b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 816.115128] env[61243]: DEBUG oslo_concurrency.lockutils [req-55811a87-7039-4ee8-a6d8-1eedee8a12f0 req-cba86aa3-45f6-46c8-81df-d2c68bb367ce service nova] Acquired lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.115412] env[61243]: DEBUG nova.network.neutron [req-55811a87-7039-4ee8-a6d8-1eedee8a12f0 req-cba86aa3-45f6-46c8-81df-d2c68bb367ce service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Refreshing network info cache for port 4cd8c409-8daa-4f53-ac5f-d2ac25de247b {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 816.116579] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:60:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '93c5b7ce-4c84-40bc-884c-b2453e0eee69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4cd8c409-8daa-4f53-ac5f-d2ac25de247b', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.124261] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Creating folder: Project (683c027e4adc4cc4a6b8747a8c8c0de9). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 816.127346] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cd4efbfc-0e03-48aa-9a23-1be52e00a15b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.138535] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Created folder: Project (683c027e4adc4cc4a6b8747a8c8c0de9) in parent group-v285636. [ 816.138725] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Creating folder: Instances. Parent ref: group-v285665. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 816.139024] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cd4e79ba-472e-435c-9e08-e5b9066a600d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.150909] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Created folder: Instances in parent group-v285665. [ 816.151115] env[61243]: DEBUG oslo.service.loopingcall [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.151308] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 816.151834] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-76205aab-ae6c-436c-a5a1-b34847f849d3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.172087] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338777, 'name': Rename_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.177411] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.177411] env[61243]: value = "task-1338780" [ 816.177411] env[61243]: _type = "Task" [ 816.177411] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.186607] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338780, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.216589] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.863s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.217267] env[61243]: ERROR nova.compute.manager [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 846ec316-a3ea-48e9-aa3c-a0abf4983e5b, please check neutron logs for more information. [ 816.217267] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Traceback (most recent call last): [ 816.217267] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 816.217267] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] self.driver.spawn(context, instance, image_meta, [ 816.217267] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 816.217267] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 816.217267] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 816.217267] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] vm_ref = self.build_virtual_machine(instance, [ 816.217267] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 816.217267] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] vif_infos = vmwarevif.get_vif_info(self._session, [ 816.217267] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 816.217614] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] for vif in network_info: [ 816.217614] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 816.217614] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] return self._sync_wrapper(fn, *args, **kwargs) [ 816.217614] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 816.217614] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] self.wait() [ 816.217614] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 816.217614] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] self[:] = self._gt.wait() [ 816.217614] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 816.217614] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] return self._exit_event.wait() [ 816.217614] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 816.217614] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] result = hub.switch() [ 816.217614] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 816.217614] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] return self.greenlet.switch() [ 816.217927] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 816.217927] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] result = function(*args, **kwargs) [ 816.217927] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 816.217927] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] return func(*args, **kwargs) [ 816.217927] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 816.217927] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] raise e [ 816.217927] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 816.217927] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] nwinfo = self.network_api.allocate_for_instance( [ 816.217927] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 816.217927] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] created_port_ids = self._update_ports_for_instance( [ 816.217927] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 816.217927] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] with excutils.save_and_reraise_exception(): [ 816.217927] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 816.218300] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] self.force_reraise() [ 816.218300] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 816.218300] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] raise self.value [ 816.218300] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 816.218300] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] updated_port = self._update_port( [ 816.218300] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 816.218300] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] _ensure_no_port_binding_failure(port) [ 816.218300] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 816.218300] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] raise exception.PortBindingFailed(port_id=port['id']) [ 816.218300] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] nova.exception.PortBindingFailed: Binding failed for port 846ec316-a3ea-48e9-aa3c-a0abf4983e5b, please check neutron logs for more information. [ 816.218300] env[61243]: ERROR nova.compute.manager [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] [ 816.218602] env[61243]: DEBUG nova.compute.utils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Binding failed for port 846ec316-a3ea-48e9-aa3c-a0abf4983e5b, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 816.219122] env[61243]: DEBUG oslo_concurrency.lockutils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.082s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.222049] env[61243]: DEBUG nova.compute.manager [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Build of instance f42f07cd-3d94-41a3-b202-ca1fa6b75da5 was re-scheduled: Binding failed for port 846ec316-a3ea-48e9-aa3c-a0abf4983e5b, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 816.222532] env[61243]: DEBUG nova.compute.manager [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 816.222763] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "refresh_cache-f42f07cd-3d94-41a3-b202-ca1fa6b75da5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.222971] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "refresh_cache-f42f07cd-3d94-41a3-b202-ca1fa6b75da5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.223167] env[61243]: DEBUG nova.network.neutron [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 816.276983] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a8946c-79e6-767c-c18b-243e2bac727c, 'name': SearchDatastore_Task, 'duration_secs': 0.012407} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.279641] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.279915] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 36e55334-8628-4dd7-a845-f4ae3d8e7ff9/36e55334-8628-4dd7-a845-f4ae3d8e7ff9.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 816.280201] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-93a72409-9c84-4577-8d63-cffaf9832705 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.287224] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 816.287224] env[61243]: value = "task-1338781" [ 816.287224] env[61243]: _type = "Task" [ 816.287224] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.295286] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338781, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.307973] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c06d0d07-1f4c-4f00-9200-4b438cc2eec9 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "86179fb0-99df-4b10-a815-c19168e9521e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.946s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.355622] env[61243]: DEBUG nova.network.neutron [req-55811a87-7039-4ee8-a6d8-1eedee8a12f0 req-cba86aa3-45f6-46c8-81df-d2c68bb367ce service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Updated VIF entry in instance network info cache for port 4cd8c409-8daa-4f53-ac5f-d2ac25de247b. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 816.355999] env[61243]: DEBUG nova.network.neutron [req-55811a87-7039-4ee8-a6d8-1eedee8a12f0 req-cba86aa3-45f6-46c8-81df-d2c68bb367ce service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Updating instance_info_cache with network_info: [{"id": "4cd8c409-8daa-4f53-ac5f-d2ac25de247b", "address": "fa:16:3e:11:60:2d", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4cd8c409-8d", "ovs_interfaceid": "4cd8c409-8daa-4f53-ac5f-d2ac25de247b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.374522] env[61243]: DEBUG nova.compute.manager [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 816.405718] env[61243]: DEBUG nova.virt.hardware [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 816.405718] env[61243]: DEBUG nova.virt.hardware [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 816.405912] env[61243]: DEBUG nova.virt.hardware [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 816.406722] env[61243]: DEBUG nova.virt.hardware [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 816.406722] env[61243]: DEBUG nova.virt.hardware [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 816.406722] env[61243]: DEBUG nova.virt.hardware [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 816.406722] env[61243]: DEBUG nova.virt.hardware [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 816.406900] env[61243]: DEBUG nova.virt.hardware [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 816.406960] env[61243]: DEBUG nova.virt.hardware [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 816.407148] env[61243]: DEBUG nova.virt.hardware [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 816.407366] env[61243]: DEBUG nova.virt.hardware [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 816.408270] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602ddb87-60ab-4afe-8a67-2dd19b3cc08b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.416481] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7f5ef4-4cbc-4dd6-84f8-d731ee875688 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.433296] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Instance VIF info [] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.443093] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Creating folder: Project (fad91f3d44114ddd9ece942ef4c93849). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 816.444026] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ddd49ea4-b337-464d-92dc-23d835f74958 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.454284] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Created folder: Project (fad91f3d44114ddd9ece942ef4c93849) in parent group-v285636. [ 816.454489] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Creating folder: Instances. Parent ref: group-v285668. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 816.454880] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ef1c739c-4d6e-487c-a417-47c689c3ac3e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.464723] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Created folder: Instances in parent group-v285668. [ 816.464923] env[61243]: DEBUG oslo.service.loopingcall [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.465179] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 816.465396] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-117f704d-ba9e-4138-a7b4-e7e4f4838bb4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.482856] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.482856] env[61243]: value = "task-1338784" [ 816.482856] env[61243]: _type = "Task" [ 816.482856] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.491687] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338784, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.659863] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338777, 'name': Rename_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.688342] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338780, 'name': CreateVM_Task, 'duration_secs': 0.376865} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.688550] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 816.689300] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.689472] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.689863] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 816.690163] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46f71b41-ffce-44f9-bfb6-6357c8c69893 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.695150] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 816.695150] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f7f7a5-7dcb-be7f-f5a2-81890033cffa" [ 816.695150] env[61243]: _type = "Task" [ 816.695150] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.703451] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f7f7a5-7dcb-be7f-f5a2-81890033cffa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.748547] env[61243]: DEBUG nova.network.neutron [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.801674] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338781, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.811022] env[61243]: DEBUG nova.compute.manager [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 816.847729] env[61243]: DEBUG nova.network.neutron [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.862025] env[61243]: DEBUG oslo_concurrency.lockutils [req-55811a87-7039-4ee8-a6d8-1eedee8a12f0 req-cba86aa3-45f6-46c8-81df-d2c68bb367ce service nova] Releasing lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.999772] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338784, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.062046] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05088fa-c5d8-4d90-a429-b38ae3323f96 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.071177] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14529d8-6c58-4941-b7e8-e8a2fde2b921 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.105768] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6a2373-9f2d-4685-bac4-5b350e2424f5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.114878] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc49f82-d0ab-42f0-a256-11b2f664168b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.129370] env[61243]: DEBUG nova.compute.provider_tree [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 817.157702] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338777, 'name': Rename_Task, 'duration_secs': 1.070948} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.157920] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 817.158242] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-56edb908-b88f-4666-9f4c-f31bd0979c35 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.164623] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 817.164623] env[61243]: value = "task-1338785" [ 817.164623] env[61243]: _type = "Task" [ 817.164623] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.172215] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338785, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.205171] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f7f7a5-7dcb-be7f-f5a2-81890033cffa, 'name': SearchDatastore_Task, 'duration_secs': 0.057963} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.205481] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.205713] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 817.205944] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.206155] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.206327] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 817.206587] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b83fb1e8-7607-42d4-8758-eada8fcd555c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.213819] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 817.213996] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 817.214716] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82860aa3-1a87-4983-b0dc-ebca83f1fcba {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.220056] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 817.220056] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d6e1e2-3dc0-5d9e-88dd-8ca2840d55ff" [ 817.220056] env[61243]: _type = "Task" [ 817.220056] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.228649] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d6e1e2-3dc0-5d9e-88dd-8ca2840d55ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.297795] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338781, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.805736} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.298131] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 36e55334-8628-4dd7-a845-f4ae3d8e7ff9/36e55334-8628-4dd7-a845-f4ae3d8e7ff9.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 817.298361] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 817.298604] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-516e8ce0-8fb0-4ab6-bbd9-28bb00fddd7a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.304824] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 817.304824] env[61243]: value = "task-1338786" [ 817.304824] env[61243]: _type = "Task" [ 817.304824] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.312146] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338786, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.329918] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.352768] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "refresh_cache-f42f07cd-3d94-41a3-b202-ca1fa6b75da5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.353059] env[61243]: DEBUG nova.compute.manager [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 817.353265] env[61243]: DEBUG nova.compute.manager [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 817.353448] env[61243]: DEBUG nova.network.neutron [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 817.376356] env[61243]: DEBUG nova.network.neutron [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.482472] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "5daccb48-b9bf-423a-ab1a-5aa26d8bec87" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.482686] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "5daccb48-b9bf-423a-ab1a-5aa26d8bec87" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.494614] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338784, 'name': CreateVM_Task, 'duration_secs': 0.698476} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.494806] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 817.495258] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.495426] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.495746] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.496018] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc90d991-636f-4546-9014-37bb5a112032 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.500733] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 817.500733] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52aec5cd-8d47-35a3-668b-35c924fc2cfc" [ 817.500733] env[61243]: _type = "Task" [ 817.500733] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.509053] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52aec5cd-8d47-35a3-668b-35c924fc2cfc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.653315] env[61243]: ERROR nova.scheduler.client.report [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [req-c5463276-f817-477b-9b58-b4133f8e9337] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c5463276-f817-477b-9b58-b4133f8e9337"}]}: nova.exception.PortBindingFailed: Binding failed for port 047ef0cd-631f-4b18-8690-72576798bb0a, please check neutron logs for more information. [ 817.670607] env[61243]: DEBUG nova.scheduler.client.report [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 817.678153] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338785, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.685244] env[61243]: DEBUG nova.scheduler.client.report [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 817.685467] env[61243]: DEBUG nova.compute.provider_tree [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 817.697061] env[61243]: DEBUG nova.scheduler.client.report [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 817.715726] env[61243]: DEBUG nova.scheduler.client.report [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 817.730887] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d6e1e2-3dc0-5d9e-88dd-8ca2840d55ff, 'name': SearchDatastore_Task, 'duration_secs': 0.011526} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.730887] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-443149e8-1c3f-4e72-ae27-d403f4a1fb3b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.739034] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 817.739034] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e4fea4-1078-7277-3988-e59aff4e5283" [ 817.739034] env[61243]: _type = "Task" [ 817.739034] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.746856] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e4fea4-1078-7277-3988-e59aff4e5283, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.814361] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338786, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06907} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.814638] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 817.815458] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db63dd5e-574c-4c74-b4af-8b659dd9e873 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.839321] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 36e55334-8628-4dd7-a845-f4ae3d8e7ff9/36e55334-8628-4dd7-a845-f4ae3d8e7ff9.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 817.841928] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bbc2d4a0-e2a1-47fa-89a0-dc252722fa96 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.861034] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 817.861034] env[61243]: value = "task-1338787" [ 817.861034] env[61243]: _type = "Task" [ 817.861034] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.871038] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338787, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.881838] env[61243]: DEBUG nova.network.neutron [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.006930] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4154fbc-e75d-40c1-8b8d-797cf1f2bac9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.012944] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52aec5cd-8d47-35a3-668b-35c924fc2cfc, 'name': SearchDatastore_Task, 'duration_secs': 0.011417} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.013693] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.013847] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 818.014082] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.017316] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f04c3c-56cb-4729-a9ab-c6d4084a4346 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.046153] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ae73a7-463e-42a9-a983-dcb0689f259e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.054031] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-870798ee-59d7-444d-bdcb-5cfcdbfe08aa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.066952] env[61243]: DEBUG nova.compute.provider_tree [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 818.174900] env[61243]: DEBUG oslo_vmware.api [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338785, 'name': PowerOnVM_Task, 'duration_secs': 0.524965} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.175200] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 818.175443] env[61243]: INFO nova.compute.manager [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Took 9.09 seconds to spawn the instance on the hypervisor. [ 818.175657] env[61243]: DEBUG nova.compute.manager [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 818.176450] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2def4c-8388-4f89-bf6e-ea88aaa70e81 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.248689] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e4fea4-1078-7277-3988-e59aff4e5283, 'name': SearchDatastore_Task, 'duration_secs': 0.045087} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.248689] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.248858] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00/7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 818.249128] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.249325] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 818.249535] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fad84c25-7dca-4084-ada7-ad6670fe91ce {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.251533] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7c2d974d-3fe7-4016-9b3a-b808d9433262 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.260724] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 818.260724] env[61243]: value = "task-1338788" [ 818.260724] env[61243]: _type = "Task" [ 818.260724] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.265254] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 818.265455] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 818.268787] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-895eddbd-747a-4b00-a15d-69e5fd558929 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.270790] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338788, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.274113] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 818.274113] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52769f41-0687-2d49-917c-fbfe357c5725" [ 818.274113] env[61243]: _type = "Task" [ 818.274113] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.281601] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52769f41-0687-2d49-917c-fbfe357c5725, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.371326] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338787, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.384051] env[61243]: INFO nova.compute.manager [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: f42f07cd-3d94-41a3-b202-ca1fa6b75da5] Took 1.03 seconds to deallocate network for instance. [ 818.598627] env[61243]: DEBUG nova.scheduler.client.report [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 83 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 818.598887] env[61243]: DEBUG nova.compute.provider_tree [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 83 to 84 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 818.599098] env[61243]: DEBUG nova.compute.provider_tree [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 818.693076] env[61243]: INFO nova.compute.manager [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Took 38.75 seconds to build instance. [ 818.770774] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338788, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.783693] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52769f41-0687-2d49-917c-fbfe357c5725, 'name': SearchDatastore_Task, 'duration_secs': 0.047187} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.784891] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c491ba34-ff61-45e1-8fc5-2c56a324adcb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.789509] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 818.789509] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524a0e74-0820-efce-e124-fb0bbba397a2" [ 818.789509] env[61243]: _type = "Task" [ 818.789509] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.797212] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524a0e74-0820-efce-e124-fb0bbba397a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.875228] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338787, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.104244] env[61243]: DEBUG oslo_concurrency.lockutils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.885s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.104714] env[61243]: ERROR nova.compute.manager [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 047ef0cd-631f-4b18-8690-72576798bb0a, please check neutron logs for more information. [ 819.104714] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Traceback (most recent call last): [ 819.104714] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 819.104714] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] self.driver.spawn(context, instance, image_meta, [ 819.104714] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 819.104714] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] self._vmops.spawn(context, instance, image_meta, injected_files, [ 819.104714] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 819.104714] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] vm_ref = self.build_virtual_machine(instance, [ 819.104714] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 819.104714] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] vif_infos = vmwarevif.get_vif_info(self._session, [ 819.104714] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 819.105040] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] for vif in network_info: [ 819.105040] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 819.105040] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] return self._sync_wrapper(fn, *args, **kwargs) [ 819.105040] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 819.105040] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] self.wait() [ 819.105040] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 819.105040] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] self[:] = self._gt.wait() [ 819.105040] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 819.105040] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] return self._exit_event.wait() [ 819.105040] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 819.105040] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] result = hub.switch() [ 819.105040] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 819.105040] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] return self.greenlet.switch() [ 819.105421] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 819.105421] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] result = function(*args, **kwargs) [ 819.105421] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 819.105421] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] return func(*args, **kwargs) [ 819.105421] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 819.105421] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] raise e [ 819.105421] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 819.105421] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] nwinfo = self.network_api.allocate_for_instance( [ 819.105421] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 819.105421] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] created_port_ids = self._update_ports_for_instance( [ 819.105421] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 819.105421] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] with excutils.save_and_reraise_exception(): [ 819.105421] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 819.105805] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] self.force_reraise() [ 819.105805] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 819.105805] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] raise self.value [ 819.105805] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 819.105805] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] updated_port = self._update_port( [ 819.105805] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 819.105805] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] _ensure_no_port_binding_failure(port) [ 819.105805] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 819.105805] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] raise exception.PortBindingFailed(port_id=port['id']) [ 819.105805] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] nova.exception.PortBindingFailed: Binding failed for port 047ef0cd-631f-4b18-8690-72576798bb0a, please check neutron logs for more information. [ 819.105805] env[61243]: ERROR nova.compute.manager [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] [ 819.106123] env[61243]: DEBUG nova.compute.utils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Binding failed for port 047ef0cd-631f-4b18-8690-72576798bb0a, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 819.106899] env[61243]: DEBUG oslo_concurrency.lockutils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.339s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.108411] env[61243]: INFO nova.compute.claims [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.111235] env[61243]: DEBUG nova.compute.manager [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Build of instance 6bd8ea50-e23f-48e7-9979-019e7a694c66 was re-scheduled: Binding failed for port 047ef0cd-631f-4b18-8690-72576798bb0a, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 819.111710] env[61243]: DEBUG nova.compute.manager [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 819.111949] env[61243]: DEBUG oslo_concurrency.lockutils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Acquiring lock "refresh_cache-6bd8ea50-e23f-48e7-9979-019e7a694c66" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.112117] env[61243]: DEBUG oslo_concurrency.lockutils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Acquired lock "refresh_cache-6bd8ea50-e23f-48e7-9979-019e7a694c66" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.112281] env[61243]: DEBUG nova.network.neutron [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 819.194327] env[61243]: DEBUG oslo_concurrency.lockutils [None req-602132ac-c3d2-4359-8474-08994f9ad258 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.840s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.271486] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338788, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.875565} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.271698] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00/7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 819.271909] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 819.272169] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c921d11d-b60f-4c92-9fb5-ab5b0a2dadca {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.278556] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 819.278556] env[61243]: value = "task-1338789" [ 819.278556] env[61243]: _type = "Task" [ 819.278556] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.287074] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338789, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.299039] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524a0e74-0820-efce-e124-fb0bbba397a2, 'name': SearchDatastore_Task, 'duration_secs': 0.008018} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.299335] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.299626] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 6cee3dd1-c13d-427d-889c-6cd3ae388a31/6cee3dd1-c13d-427d-889c-6cd3ae388a31.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 819.299914] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a0a4fc4-58a9-4e1b-8687-86db51584e44 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.306540] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 819.306540] env[61243]: value = "task-1338790" [ 819.306540] env[61243]: _type = "Task" [ 819.306540] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.313937] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338790, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.372522] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338787, 'name': ReconfigVM_Task, 'duration_secs': 1.023405} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.372840] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 36e55334-8628-4dd7-a845-f4ae3d8e7ff9/36e55334-8628-4dd7-a845-f4ae3d8e7ff9.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.373545] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-50ffa975-911d-4bde-b47c-0dffd4115798 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.379475] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 819.379475] env[61243]: value = "task-1338791" [ 819.379475] env[61243]: _type = "Task" [ 819.379475] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.387508] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338791, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.421409] env[61243]: INFO nova.scheduler.client.report [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleted allocations for instance f42f07cd-3d94-41a3-b202-ca1fa6b75da5 [ 819.639776] env[61243]: DEBUG nova.network.neutron [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 819.698022] env[61243]: DEBUG nova.compute.manager [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 819.788931] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338789, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076224} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.791135] env[61243]: DEBUG nova.network.neutron [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.791806] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 819.795168] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d8df33-e226-4cc8-acf8-7f2e379f8cae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.819587] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00/7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.824297] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2e42bd4-acec-497b-86b7-68253bf57aca {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.839190] env[61243]: DEBUG nova.compute.manager [req-ed93ff15-1e68-4733-a4d3-dc1b9c985ab9 req-0be5b719-342f-46dc-8c6d-187a29987bb0 service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Received event network-changed-af7195a1-9a59-4b3a-b62d-7e907c94ce7b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.839404] env[61243]: DEBUG nova.compute.manager [req-ed93ff15-1e68-4733-a4d3-dc1b9c985ab9 req-0be5b719-342f-46dc-8c6d-187a29987bb0 service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Refreshing instance network info cache due to event network-changed-af7195a1-9a59-4b3a-b62d-7e907c94ce7b. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 819.839623] env[61243]: DEBUG oslo_concurrency.lockutils [req-ed93ff15-1e68-4733-a4d3-dc1b9c985ab9 req-0be5b719-342f-46dc-8c6d-187a29987bb0 service nova] Acquiring lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.839770] env[61243]: DEBUG oslo_concurrency.lockutils [req-ed93ff15-1e68-4733-a4d3-dc1b9c985ab9 req-0be5b719-342f-46dc-8c6d-187a29987bb0 service nova] Acquired lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.839929] env[61243]: DEBUG nova.network.neutron [req-ed93ff15-1e68-4733-a4d3-dc1b9c985ab9 req-0be5b719-342f-46dc-8c6d-187a29987bb0 service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Refreshing network info cache for port af7195a1-9a59-4b3a-b62d-7e907c94ce7b {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 819.849222] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338790, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.852499] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 819.852499] env[61243]: value = "task-1338792" [ 819.852499] env[61243]: _type = "Task" [ 819.852499] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.859575] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338792, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.888869] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338791, 'name': Rename_Task, 'duration_secs': 0.189641} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.889635] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 819.889896] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-18ce996b-aeaf-46e2-83ee-71eec5f52257 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.896030] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 819.896030] env[61243]: value = "task-1338793" [ 819.896030] env[61243]: _type = "Task" [ 819.896030] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.904686] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338793, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.929297] env[61243]: DEBUG oslo_concurrency.lockutils [None req-12b81a50-5ba7-48c2-80af-466e69c08666 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "f42f07cd-3d94-41a3-b202-ca1fa6b75da5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.043s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.216413] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.294736] env[61243]: DEBUG oslo_concurrency.lockutils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Releasing lock "refresh_cache-6bd8ea50-e23f-48e7-9979-019e7a694c66" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.294978] env[61243]: DEBUG nova.compute.manager [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 820.295200] env[61243]: DEBUG nova.compute.manager [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 820.295386] env[61243]: DEBUG nova.network.neutron [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 820.319132] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338790, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.361532] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338792, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.364890] env[61243]: DEBUG nova.network.neutron [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.393141] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbab2cf4-302a-41de-966d-5bda837781b8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.408416] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338793, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.409497] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e98b6ac5-5d64-49ec-9aca-219342fd4ac8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.448544] env[61243]: DEBUG nova.compute.manager [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 820.455978] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-137f6b4f-0aba-4bc0-844f-4c64c75e8167 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.467542] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f9eb19-b3af-4527-aec0-e525488bc92f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.481398] env[61243]: DEBUG nova.compute.provider_tree [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.795099] env[61243]: DEBUG nova.network.neutron [req-ed93ff15-1e68-4733-a4d3-dc1b9c985ab9 req-0be5b719-342f-46dc-8c6d-187a29987bb0 service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Updated VIF entry in instance network info cache for port af7195a1-9a59-4b3a-b62d-7e907c94ce7b. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 820.795476] env[61243]: DEBUG nova.network.neutron [req-ed93ff15-1e68-4733-a4d3-dc1b9c985ab9 req-0be5b719-342f-46dc-8c6d-187a29987bb0 service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Updating instance_info_cache with network_info: [{"id": "af7195a1-9a59-4b3a-b62d-7e907c94ce7b", "address": "fa:16:3e:ee:b7:e1", "network": {"id": "c6d4943d-4a8d-41b4-8023-c4f1549d312d", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1045259606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8549626a07e941e1ab03452355a0f116", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf7195a1-9a", "ovs_interfaceid": "af7195a1-9a59-4b3a-b62d-7e907c94ce7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.823258] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338790, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.499214} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.823536] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 6cee3dd1-c13d-427d-889c-6cd3ae388a31/6cee3dd1-c13d-427d-889c-6cd3ae388a31.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 820.823752] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 820.824020] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2d091896-c4ae-496b-8ed7-93f915d5b712 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.829563] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 820.829563] env[61243]: value = "task-1338794" [ 820.829563] env[61243]: _type = "Task" [ 820.829563] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.837757] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338794, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.861751] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338792, 'name': ReconfigVM_Task, 'duration_secs': 0.95528} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.862047] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00/7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 820.862689] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1537f3e5-a7fd-426b-bd06-8e8866d7e967 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.868553] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 820.868553] env[61243]: value = "task-1338795" [ 820.868553] env[61243]: _type = "Task" [ 820.868553] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.869103] env[61243]: DEBUG nova.network.neutron [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.878990] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338795, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.906305] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338793, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.972889] env[61243]: DEBUG oslo_concurrency.lockutils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.985269] env[61243]: DEBUG nova.scheduler.client.report [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.298060] env[61243]: DEBUG oslo_concurrency.lockutils [req-ed93ff15-1e68-4733-a4d3-dc1b9c985ab9 req-0be5b719-342f-46dc-8c6d-187a29987bb0 service nova] Releasing lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.340834] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338794, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079247} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.341154] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 821.342060] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664feec2-a117-4a8b-adb3-96f4dbf10d49 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.361295] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 6cee3dd1-c13d-427d-889c-6cd3ae388a31/6cee3dd1-c13d-427d-889c-6cd3ae388a31.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 821.361575] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4025bf1e-fc68-4828-ab18-7baa3f9f0ada {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.376854] env[61243]: INFO nova.compute.manager [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] [instance: 6bd8ea50-e23f-48e7-9979-019e7a694c66] Took 1.08 seconds to deallocate network for instance. [ 821.384883] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "ca230425-aaa7-4c32-81d3-379cbce0d41b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.385176] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "ca230425-aaa7-4c32-81d3-379cbce0d41b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.390029] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338795, 'name': Rename_Task, 'duration_secs': 0.175766} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.391180] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 821.391699] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 821.391699] env[61243]: value = "task-1338796" [ 821.391699] env[61243]: _type = "Task" [ 821.391699] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.391884] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a06e19a-ab3a-4615-a70c-42b2a1ead337 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.404617] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338796, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.404962] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 821.404962] env[61243]: value = "task-1338797" [ 821.404962] env[61243]: _type = "Task" [ 821.404962] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.416558] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338797, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.419878] env[61243]: DEBUG oslo_vmware.api [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338793, 'name': PowerOnVM_Task, 'duration_secs': 1.063237} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.419878] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 821.419998] env[61243]: INFO nova.compute.manager [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Took 9.89 seconds to spawn the instance on the hypervisor. [ 821.420222] env[61243]: DEBUG nova.compute.manager [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 821.421516] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9565475-6d39-4232-baab-628703da7eb6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.490616] env[61243]: DEBUG oslo_concurrency.lockutils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.384s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.491129] env[61243]: DEBUG nova.compute.manager [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 821.494103] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.147s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.903339] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338796, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.914828] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338797, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.938565] env[61243]: INFO nova.compute.manager [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Took 40.10 seconds to build instance. [ 821.998946] env[61243]: DEBUG nova.compute.utils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.006612] env[61243]: DEBUG nova.compute.manager [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 822.007815] env[61243]: DEBUG nova.network.neutron [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 822.060552] env[61243]: DEBUG nova.policy [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7847c54c692c4a8f9ddf93aa32e88690', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4fcef2317fdf47f8ac820aeef1765fef', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 822.273803] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47770c3c-6121-4802-ad07-7bc98d678a86 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.281413] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71804d7f-c3da-478a-81cf-7893690c7041 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.317380] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc323c43-af7e-494c-a6c8-9fa6f5196616 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.320327] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f0099d-6966-45e6-9b49-01a75a715d68 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.329677] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca54411-73ea-4e36-a486-73e5207ca928 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.333398] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7112303a-e8ec-4b0b-b369-3d1c2059da97 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Suspending the VM {{(pid=61243) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 822.334137] env[61243]: DEBUG nova.network.neutron [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Successfully created port: bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 822.335810] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-b74e30df-d258-4491-a046-4c07e0fe0f11 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.346646] env[61243]: DEBUG nova.compute.provider_tree [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.348977] env[61243]: DEBUG oslo_vmware.api [None req-7112303a-e8ec-4b0b-b369-3d1c2059da97 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 822.348977] env[61243]: value = "task-1338798" [ 822.348977] env[61243]: _type = "Task" [ 822.348977] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.356582] env[61243]: DEBUG oslo_vmware.api [None req-7112303a-e8ec-4b0b-b369-3d1c2059da97 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338798, 'name': SuspendVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.402781] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338796, 'name': ReconfigVM_Task, 'duration_secs': 0.612584} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.404028] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 6cee3dd1-c13d-427d-889c-6cd3ae388a31/6cee3dd1-c13d-427d-889c-6cd3ae388a31.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 822.404028] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f8da757a-4614-4567-956d-7dfd7305061f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.411302] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 822.411302] env[61243]: value = "task-1338799" [ 822.411302] env[61243]: _type = "Task" [ 822.411302] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.412248] env[61243]: INFO nova.scheduler.client.report [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Deleted allocations for instance 6bd8ea50-e23f-48e7-9979-019e7a694c66 [ 822.426327] env[61243]: DEBUG oslo_vmware.api [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338797, 'name': PowerOnVM_Task, 'duration_secs': 0.735933} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.427062] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 822.427312] env[61243]: INFO nova.compute.manager [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Took 8.50 seconds to spawn the instance on the hypervisor. [ 822.427634] env[61243]: DEBUG nova.compute.manager [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 822.428392] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74000952-b2d4-423d-9ed7-d5e6753cfab1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.434700] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338799, 'name': Rename_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.442358] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e5d8228-094a-499d-9e0a-b4c6175d12a4 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "36e55334-8628-4dd7-a845-f4ae3d8e7ff9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.357s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.509732] env[61243]: DEBUG nova.compute.manager [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 822.851145] env[61243]: DEBUG nova.scheduler.client.report [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 822.863429] env[61243]: DEBUG oslo_vmware.api [None req-7112303a-e8ec-4b0b-b369-3d1c2059da97 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338798, 'name': SuspendVM_Task} progress is 62%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.928378] env[61243]: DEBUG oslo_concurrency.lockutils [None req-61e67a04-d8fd-42d7-85ad-0d87f6c6ea3d tempest-VolumesAdminNegativeTest-99707993 tempest-VolumesAdminNegativeTest-99707993-project-member] Lock "6bd8ea50-e23f-48e7-9979-019e7a694c66" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.449s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.928687] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338799, 'name': Rename_Task, 'duration_secs': 0.269067} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.930623] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 822.931380] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-58db55f9-8000-4b6e-aa33-11b973c28a63 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.937376] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 822.937376] env[61243]: value = "task-1338800" [ 822.937376] env[61243]: _type = "Task" [ 822.937376] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.950147] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338800, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.951012] env[61243]: DEBUG nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 822.955830] env[61243]: INFO nova.compute.manager [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Took 39.08 seconds to build instance. [ 823.362046] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.866s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.362046] env[61243]: ERROR nova.compute.manager [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cff38ea8-81bd-4bd5-b704-b61286d84199, please check neutron logs for more information. [ 823.362046] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Traceback (most recent call last): [ 823.362046] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 823.362046] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] self.driver.spawn(context, instance, image_meta, [ 823.362046] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 823.362046] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 823.362046] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 823.362046] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] vm_ref = self.build_virtual_machine(instance, [ 823.362391] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 823.362391] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] vif_infos = vmwarevif.get_vif_info(self._session, [ 823.362391] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 823.362391] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] for vif in network_info: [ 823.362391] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 823.362391] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] return self._sync_wrapper(fn, *args, **kwargs) [ 823.362391] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 823.362391] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] self.wait() [ 823.362391] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 823.362391] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] self[:] = self._gt.wait() [ 823.362391] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 823.362391] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] return self._exit_event.wait() [ 823.362391] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 823.362696] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] current.throw(*self._exc) [ 823.362696] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 823.362696] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] result = function(*args, **kwargs) [ 823.362696] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 823.362696] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] return func(*args, **kwargs) [ 823.362696] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 823.362696] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] raise e [ 823.362696] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 823.362696] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] nwinfo = self.network_api.allocate_for_instance( [ 823.362696] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 823.362696] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] created_port_ids = self._update_ports_for_instance( [ 823.362696] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 823.362696] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] with excutils.save_and_reraise_exception(): [ 823.362984] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 823.362984] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] self.force_reraise() [ 823.362984] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 823.362984] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] raise self.value [ 823.362984] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 823.362984] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] updated_port = self._update_port( [ 823.362984] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 823.362984] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] _ensure_no_port_binding_failure(port) [ 823.362984] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 823.362984] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] raise exception.PortBindingFailed(port_id=port['id']) [ 823.362984] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] nova.exception.PortBindingFailed: Binding failed for port cff38ea8-81bd-4bd5-b704-b61286d84199, please check neutron logs for more information. [ 823.362984] env[61243]: ERROR nova.compute.manager [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] [ 823.363404] env[61243]: DEBUG nova.compute.utils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Binding failed for port cff38ea8-81bd-4bd5-b704-b61286d84199, please check neutron logs for more information. {{(pid=61243) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 823.364729] env[61243]: DEBUG nova.compute.manager [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Build of instance e9b2e5d1-e75e-4346-a0af-c0dafac0692e was re-scheduled: Binding failed for port cff38ea8-81bd-4bd5-b704-b61286d84199, please check neutron logs for more information. {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 823.368128] env[61243]: DEBUG nova.compute.manager [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Unplugging VIFs for instance {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 823.368128] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Acquiring lock "refresh_cache-e9b2e5d1-e75e-4346-a0af-c0dafac0692e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.368128] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Acquired lock "refresh_cache-e9b2e5d1-e75e-4346-a0af-c0dafac0692e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.368128] env[61243]: DEBUG nova.network.neutron [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 823.368128] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.109s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.369885] env[61243]: INFO nova.compute.claims [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.379917] env[61243]: DEBUG oslo_vmware.api [None req-7112303a-e8ec-4b0b-b369-3d1c2059da97 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338798, 'name': SuspendVM_Task} progress is 62%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.435025] env[61243]: DEBUG nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 823.450771] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338800, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.461805] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3a0205b6-f78f-428a-94c0-f7b45e9061b8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.822s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.487358] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.521979] env[61243]: DEBUG nova.compute.manager [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 823.575201] env[61243]: DEBUG nova.virt.hardware [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 823.575499] env[61243]: DEBUG nova.virt.hardware [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 823.575655] env[61243]: DEBUG nova.virt.hardware [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 823.575832] env[61243]: DEBUG nova.virt.hardware [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 823.575974] env[61243]: DEBUG nova.virt.hardware [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 823.576190] env[61243]: DEBUG nova.virt.hardware [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 823.576360] env[61243]: DEBUG nova.virt.hardware [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 823.576524] env[61243]: DEBUG nova.virt.hardware [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 823.576684] env[61243]: DEBUG nova.virt.hardware [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 823.576841] env[61243]: DEBUG nova.virt.hardware [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 823.577020] env[61243]: DEBUG nova.virt.hardware [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 823.579394] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d5cb52-8df9-43fb-8998-901b5bf5c668 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.585673] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b2faf8-1f36-4e8a-a6ea-e06e94efdc40 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.605143] env[61243]: DEBUG nova.compute.manager [req-6de6d78e-6cf1-44bd-a91a-6433566c29b5 req-9f80109b-8b9a-44c4-a34a-2b2bf5b33dcf service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Received event network-changed-4cd8c409-8daa-4f53-ac5f-d2ac25de247b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.606071] env[61243]: DEBUG nova.compute.manager [req-6de6d78e-6cf1-44bd-a91a-6433566c29b5 req-9f80109b-8b9a-44c4-a34a-2b2bf5b33dcf service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Refreshing instance network info cache due to event network-changed-4cd8c409-8daa-4f53-ac5f-d2ac25de247b. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 823.606071] env[61243]: DEBUG oslo_concurrency.lockutils [req-6de6d78e-6cf1-44bd-a91a-6433566c29b5 req-9f80109b-8b9a-44c4-a34a-2b2bf5b33dcf service nova] Acquiring lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.606071] env[61243]: DEBUG oslo_concurrency.lockutils [req-6de6d78e-6cf1-44bd-a91a-6433566c29b5 req-9f80109b-8b9a-44c4-a34a-2b2bf5b33dcf service nova] Acquired lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.606071] env[61243]: DEBUG nova.network.neutron [req-6de6d78e-6cf1-44bd-a91a-6433566c29b5 req-9f80109b-8b9a-44c4-a34a-2b2bf5b33dcf service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Refreshing network info cache for port 4cd8c409-8daa-4f53-ac5f-d2ac25de247b {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 823.847461] env[61243]: DEBUG nova.network.neutron [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Successfully updated port: bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 823.866632] env[61243]: DEBUG oslo_vmware.api [None req-7112303a-e8ec-4b0b-b369-3d1c2059da97 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338798, 'name': SuspendVM_Task, 'duration_secs': 1.177864} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.866934] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7112303a-e8ec-4b0b-b369-3d1c2059da97 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Suspended the VM {{(pid=61243) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 823.867134] env[61243]: DEBUG nova.compute.manager [None req-7112303a-e8ec-4b0b-b369-3d1c2059da97 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.868185] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9312ff1c-4e82-4c69-a13f-c4ce1814eef2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.894478] env[61243]: DEBUG nova.network.neutron [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.953172] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338800, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.966174] env[61243]: DEBUG nova.compute.manager [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 823.969649] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.974550] env[61243]: DEBUG nova.network.neutron [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.349125] env[61243]: DEBUG oslo_concurrency.lockutils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "refresh_cache-894eb798-6c9e-47cb-8eb5-4610fe184bc5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.349125] env[61243]: DEBUG oslo_concurrency.lockutils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquired lock "refresh_cache-894eb798-6c9e-47cb-8eb5-4610fe184bc5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.349379] env[61243]: DEBUG nova.network.neutron [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 824.354123] env[61243]: DEBUG nova.network.neutron [req-6de6d78e-6cf1-44bd-a91a-6433566c29b5 req-9f80109b-8b9a-44c4-a34a-2b2bf5b33dcf service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Updated VIF entry in instance network info cache for port 4cd8c409-8daa-4f53-ac5f-d2ac25de247b. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 824.354489] env[61243]: DEBUG nova.network.neutron [req-6de6d78e-6cf1-44bd-a91a-6433566c29b5 req-9f80109b-8b9a-44c4-a34a-2b2bf5b33dcf service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Updating instance_info_cache with network_info: [{"id": "4cd8c409-8daa-4f53-ac5f-d2ac25de247b", "address": "fa:16:3e:11:60:2d", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4cd8c409-8d", "ovs_interfaceid": "4cd8c409-8daa-4f53-ac5f-d2ac25de247b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.448560] env[61243]: DEBUG oslo_vmware.api [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338800, 'name': PowerOnVM_Task, 'duration_secs': 1.142054} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.451022] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 824.451245] env[61243]: INFO nova.compute.manager [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Took 8.08 seconds to spawn the instance on the hypervisor. [ 824.451425] env[61243]: DEBUG nova.compute.manager [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 824.453563] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4ffdc97-1b1b-4239-ba9a-e406a8717dd8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.478425] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Releasing lock "refresh_cache-e9b2e5d1-e75e-4346-a0af-c0dafac0692e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.478639] env[61243]: DEBUG nova.compute.manager [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61243) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 824.478809] env[61243]: DEBUG nova.compute.manager [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 824.478965] env[61243]: DEBUG nova.network.neutron [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 824.496610] env[61243]: DEBUG nova.network.neutron [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.504455] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.683942] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d046bb55-7658-4f32-8c9e-7a9b0a283b36 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.691818] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69eeed3f-7fd5-477f-b7f5-dcfbbabb56c9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.722336] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d482caa-5633-45b9-870f-4643bd38b175 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.729918] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c029135-8bee-4044-aa9f-9af925df8ccd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.744585] env[61243]: DEBUG nova.compute.provider_tree [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.858395] env[61243]: DEBUG oslo_concurrency.lockutils [req-6de6d78e-6cf1-44bd-a91a-6433566c29b5 req-9f80109b-8b9a-44c4-a34a-2b2bf5b33dcf service nova] Releasing lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.880862] env[61243]: DEBUG nova.network.neutron [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.981268] env[61243]: INFO nova.compute.manager [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Took 39.18 seconds to build instance. [ 825.005724] env[61243]: DEBUG nova.network.neutron [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.040306] env[61243]: DEBUG nova.network.neutron [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Updating instance_info_cache with network_info: [{"id": "bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0", "address": "fa:16:3e:f6:c2:39", "network": {"id": "8e2b6835-484b-4346-8bd1-034c53671110", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710127686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4fcef2317fdf47f8ac820aeef1765fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdfbe0d5-9b", "ovs_interfaceid": "bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.248210] env[61243]: DEBUG nova.scheduler.client.report [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.481994] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7e716e79-e36b-4b5e-aad4-73ae4fc1c45e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Lock "6cee3dd1-c13d-427d-889c-6cd3ae388a31" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.867s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.508534] env[61243]: INFO nova.compute.manager [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] [instance: e9b2e5d1-e75e-4346-a0af-c0dafac0692e] Took 1.03 seconds to deallocate network for instance. [ 825.546032] env[61243]: DEBUG oslo_concurrency.lockutils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Releasing lock "refresh_cache-894eb798-6c9e-47cb-8eb5-4610fe184bc5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.546032] env[61243]: DEBUG nova.compute.manager [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Instance network_info: |[{"id": "bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0", "address": "fa:16:3e:f6:c2:39", "network": {"id": "8e2b6835-484b-4346-8bd1-034c53671110", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710127686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4fcef2317fdf47f8ac820aeef1765fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdfbe0d5-9b", "ovs_interfaceid": "bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 825.547143] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:c2:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '415e68b4-3766-4359-afe2-f8563910d98c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.556609] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Creating folder: Project (4fcef2317fdf47f8ac820aeef1765fef). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.556609] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bad28ffd-c508-4e49-9c80-a79c6a39f0f1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.567821] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Created folder: Project (4fcef2317fdf47f8ac820aeef1765fef) in parent group-v285636. [ 825.568068] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Creating folder: Instances. Parent ref: group-v285671. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.568294] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-27238273-8ced-422b-a46e-3f195d6df652 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.577760] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Created folder: Instances in parent group-v285671. [ 825.578009] env[61243]: DEBUG oslo.service.loopingcall [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.578202] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 825.578414] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c79dc4d-25dc-4d29-8bc5-7785ad5f79b2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.603937] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.603937] env[61243]: value = "task-1338803" [ 825.603937] env[61243]: _type = "Task" [ 825.603937] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.612298] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338803, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.745463] env[61243]: DEBUG nova.compute.manager [req-a3109d53-f201-4846-9c21-05a31f919662 req-e77c6cc4-7e98-408a-bcd9-93cdd6e34087 service nova] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Received event network-vif-plugged-bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.745722] env[61243]: DEBUG oslo_concurrency.lockutils [req-a3109d53-f201-4846-9c21-05a31f919662 req-e77c6cc4-7e98-408a-bcd9-93cdd6e34087 service nova] Acquiring lock "894eb798-6c9e-47cb-8eb5-4610fe184bc5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.745904] env[61243]: DEBUG oslo_concurrency.lockutils [req-a3109d53-f201-4846-9c21-05a31f919662 req-e77c6cc4-7e98-408a-bcd9-93cdd6e34087 service nova] Lock "894eb798-6c9e-47cb-8eb5-4610fe184bc5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.746292] env[61243]: DEBUG oslo_concurrency.lockutils [req-a3109d53-f201-4846-9c21-05a31f919662 req-e77c6cc4-7e98-408a-bcd9-93cdd6e34087 service nova] Lock "894eb798-6c9e-47cb-8eb5-4610fe184bc5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.746513] env[61243]: DEBUG nova.compute.manager [req-a3109d53-f201-4846-9c21-05a31f919662 req-e77c6cc4-7e98-408a-bcd9-93cdd6e34087 service nova] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] No waiting events found dispatching network-vif-plugged-bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 825.746681] env[61243]: WARNING nova.compute.manager [req-a3109d53-f201-4846-9c21-05a31f919662 req-e77c6cc4-7e98-408a-bcd9-93cdd6e34087 service nova] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Received unexpected event network-vif-plugged-bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0 for instance with vm_state building and task_state spawning. [ 825.746891] env[61243]: DEBUG nova.compute.manager [req-a3109d53-f201-4846-9c21-05a31f919662 req-e77c6cc4-7e98-408a-bcd9-93cdd6e34087 service nova] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Received event network-changed-bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.748015] env[61243]: DEBUG nova.compute.manager [req-a3109d53-f201-4846-9c21-05a31f919662 req-e77c6cc4-7e98-408a-bcd9-93cdd6e34087 service nova] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Refreshing instance network info cache due to event network-changed-bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 825.748015] env[61243]: DEBUG oslo_concurrency.lockutils [req-a3109d53-f201-4846-9c21-05a31f919662 req-e77c6cc4-7e98-408a-bcd9-93cdd6e34087 service nova] Acquiring lock "refresh_cache-894eb798-6c9e-47cb-8eb5-4610fe184bc5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.748015] env[61243]: DEBUG oslo_concurrency.lockutils [req-a3109d53-f201-4846-9c21-05a31f919662 req-e77c6cc4-7e98-408a-bcd9-93cdd6e34087 service nova] Acquired lock "refresh_cache-894eb798-6c9e-47cb-8eb5-4610fe184bc5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.748015] env[61243]: DEBUG nova.network.neutron [req-a3109d53-f201-4846-9c21-05a31f919662 req-e77c6cc4-7e98-408a-bcd9-93cdd6e34087 service nova] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Refreshing network info cache for port bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 825.753150] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.753677] env[61243]: DEBUG nova.compute.manager [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 825.756888] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.427s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.758270] env[61243]: INFO nova.compute.claims [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 825.933008] env[61243]: DEBUG nova.compute.manager [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 825.933008] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-571a8689-0a3b-4a8c-8b56-6ccaef0ee907 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.986939] env[61243]: DEBUG nova.compute.manager [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 826.101244] env[61243]: INFO nova.compute.manager [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Rebuilding instance [ 826.116998] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338803, 'name': CreateVM_Task, 'duration_secs': 0.317991} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.117194] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 826.117852] env[61243]: DEBUG oslo_concurrency.lockutils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.118031] env[61243]: DEBUG oslo_concurrency.lockutils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.118342] env[61243]: DEBUG oslo_concurrency.lockutils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 826.118594] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38e0aeed-26e0-4856-8749-ee72f589b12d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.126594] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 826.126594] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d7dd23-4c8f-6095-e9c4-9306d07deaac" [ 826.126594] env[61243]: _type = "Task" [ 826.126594] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.138096] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d7dd23-4c8f-6095-e9c4-9306d07deaac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.153205] env[61243]: DEBUG nova.compute.manager [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 826.154015] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a467ab85-bce7-45a8-87fe-29b37934b750 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.262449] env[61243]: DEBUG nova.compute.utils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.265888] env[61243]: DEBUG nova.compute.manager [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 826.266082] env[61243]: DEBUG nova.network.neutron [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 826.367997] env[61243]: DEBUG nova.policy [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32d9e3276f974c8f87573270909076ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7651b565d1647eb948fe07faba32a41', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 826.441759] env[61243]: INFO nova.compute.manager [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] instance snapshotting [ 826.441890] env[61243]: WARNING nova.compute.manager [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 826.446406] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8cd46f-782d-4747-be63-a04afac1a62b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.466415] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c2c265-3471-4e23-b6cb-6e25c8e317c7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.509390] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.544020] env[61243]: INFO nova.scheduler.client.report [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Deleted allocations for instance e9b2e5d1-e75e-4346-a0af-c0dafac0692e [ 826.637331] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d7dd23-4c8f-6095-e9c4-9306d07deaac, 'name': SearchDatastore_Task, 'duration_secs': 0.020269} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.637710] env[61243]: DEBUG oslo_concurrency.lockutils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.638131] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.638337] env[61243]: DEBUG oslo_concurrency.lockutils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.638791] env[61243]: DEBUG oslo_concurrency.lockutils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.638791] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.639187] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-deaa7d77-c844-47af-ba04-cea89da71f86 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.648671] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.652025] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 826.652025] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ed8414a-57b2-4a76-b787-f9b33adac1b5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.657174] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 826.657174] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5287f43c-5ca1-a735-90b8-97e60678c9af" [ 826.657174] env[61243]: _type = "Task" [ 826.657174] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.665427] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 826.665686] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5287f43c-5ca1-a735-90b8-97e60678c9af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.665891] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff3db2ed-22e7-4718-8a0c-99fcc5052071 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.671601] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 826.671601] env[61243]: value = "task-1338804" [ 826.671601] env[61243]: _type = "Task" [ 826.671601] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.679160] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338804, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.705951] env[61243]: DEBUG nova.network.neutron [req-a3109d53-f201-4846-9c21-05a31f919662 req-e77c6cc4-7e98-408a-bcd9-93cdd6e34087 service nova] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Updated VIF entry in instance network info cache for port bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 826.705951] env[61243]: DEBUG nova.network.neutron [req-a3109d53-f201-4846-9c21-05a31f919662 req-e77c6cc4-7e98-408a-bcd9-93cdd6e34087 service nova] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Updating instance_info_cache with network_info: [{"id": "bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0", "address": "fa:16:3e:f6:c2:39", "network": {"id": "8e2b6835-484b-4346-8bd1-034c53671110", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710127686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4fcef2317fdf47f8ac820aeef1765fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdfbe0d5-9b", "ovs_interfaceid": "bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.768167] env[61243]: DEBUG nova.compute.manager [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 826.870502] env[61243]: DEBUG nova.network.neutron [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Successfully created port: b22c18f9-e3a1-4c8b-a26a-17c651ca16b5 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 826.978625] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Creating Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 826.979227] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-11c605cb-67f9-4bd0-a41b-5bd5ce2446eb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.992887] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 826.992887] env[61243]: value = "task-1338805" [ 826.992887] env[61243]: _type = "Task" [ 826.992887] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.008626] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338805, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.164354] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9e105f01-5768-4ce6-be92-8c5c70dd2b3b tempest-InstanceActionsTestJSON-1341290575 tempest-InstanceActionsTestJSON-1341290575-project-member] Lock "e9b2e5d1-e75e-4346-a0af-c0dafac0692e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.905s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.170197] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5287f43c-5ca1-a735-90b8-97e60678c9af, 'name': SearchDatastore_Task, 'duration_secs': 0.023402} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.172153] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d6e70df-22ff-4a75-8ad5-e451e07c82a1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.187042] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 827.187042] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]526e9d06-4b9e-9fdb-cf96-687b54a2bcab" [ 827.187042] env[61243]: _type = "Task" [ 827.187042] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.192787] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338804, 'name': PowerOffVM_Task, 'duration_secs': 0.365594} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.197556] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 827.197556] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 827.197556] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5008ac-fa8a-40c6-b174-c029d2921c0e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.203807] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]526e9d06-4b9e-9fdb-cf96-687b54a2bcab, 'name': SearchDatastore_Task, 'duration_secs': 0.010218} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.204834] env[61243]: DEBUG oslo_concurrency.lockutils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.204834] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 894eb798-6c9e-47cb-8eb5-4610fe184bc5/894eb798-6c9e-47cb-8eb5-4610fe184bc5.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 827.207929] env[61243]: DEBUG oslo_concurrency.lockutils [req-a3109d53-f201-4846-9c21-05a31f919662 req-e77c6cc4-7e98-408a-bcd9-93cdd6e34087 service nova] Releasing lock "refresh_cache-894eb798-6c9e-47cb-8eb5-4610fe184bc5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.207929] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-165cb188-ab2a-45fa-83df-8e8ae3396f28 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.211613] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 827.212811] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b2cd438a-d54d-405d-babf-97773736fd5b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.217636] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 827.217636] env[61243]: value = "task-1338806" [ 827.217636] env[61243]: _type = "Task" [ 827.217636] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.223217] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8841587d-3035-4794-a139-ccca2801829b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.230720] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338806, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.234161] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a6804c-c260-4131-ab18-6449cf95dcf2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.238352] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 827.238601] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 827.238825] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Deleting the datastore file [datastore2] 6cee3dd1-c13d-427d-889c-6cd3ae388a31 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 827.239488] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9849f6c5-1964-4349-b9ba-445bbef0ed52 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.279832] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6420761a-2989-436a-b018-b4e12c73c4a1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.283043] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 827.283043] env[61243]: value = "task-1338808" [ 827.283043] env[61243]: _type = "Task" [ 827.283043] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.291286] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f92bb1cd-2977-4a56-b00e-9a111a87cb82 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.299305] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338808, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.310283] env[61243]: DEBUG nova.compute.provider_tree [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.503412] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338805, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.557480] env[61243]: DEBUG nova.compute.manager [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 827.727130] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338806, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.781222] env[61243]: DEBUG nova.compute.manager [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 827.792798] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338808, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096586} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.792798] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 827.792885] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 827.793160] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 827.811390] env[61243]: DEBUG nova.virt.hardware [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 827.811650] env[61243]: DEBUG nova.virt.hardware [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 827.811863] env[61243]: DEBUG nova.virt.hardware [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 827.812038] env[61243]: DEBUG nova.virt.hardware [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 827.812193] env[61243]: DEBUG nova.virt.hardware [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 827.812346] env[61243]: DEBUG nova.virt.hardware [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 827.812628] env[61243]: DEBUG nova.virt.hardware [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 827.812833] env[61243]: DEBUG nova.virt.hardware [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 827.813020] env[61243]: DEBUG nova.virt.hardware [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 827.813198] env[61243]: DEBUG nova.virt.hardware [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 827.813424] env[61243]: DEBUG nova.virt.hardware [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.814660] env[61243]: DEBUG nova.scheduler.client.report [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 827.818414] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef073653-6211-4f79-b594-096fc41e7a40 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.827400] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7783899f-dcb8-425b-b085-b4dc9a2701ca {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.006970] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338805, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.083972] env[61243]: DEBUG oslo_concurrency.lockutils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.235625] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338806, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510978} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.235625] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 894eb798-6c9e-47cb-8eb5-4610fe184bc5/894eb798-6c9e-47cb-8eb5-4610fe184bc5.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 828.235625] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.235625] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ebe98a57-366d-499d-a741-c8b4e6fc126a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.238592] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 828.238592] env[61243]: value = "task-1338809" [ 828.238592] env[61243]: _type = "Task" [ 828.238592] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.246325] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338809, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.323745] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.567s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.324340] env[61243]: DEBUG nova.compute.manager [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 828.327726] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.111s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.330299] env[61243]: INFO nova.compute.claims [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.508073] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338805, 'name': CreateSnapshot_Task, 'duration_secs': 1.484063} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.508425] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Created Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 828.509578] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d910d012-73dd-499c-b6f2-70f8204e865b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.750369] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338809, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083925} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.750651] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 828.757426] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e3d8bde-87fe-4840-ab1f-9cbed4246bef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.788274] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 894eb798-6c9e-47cb-8eb5-4610fe184bc5/894eb798-6c9e-47cb-8eb5-4610fe184bc5.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 828.788274] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d98c6981-58fc-4a44-a43f-3d0a6a85a62b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.809649] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 828.809649] env[61243]: value = "task-1338810" [ 828.809649] env[61243]: _type = "Task" [ 828.809649] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.821776] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338810, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.833877] env[61243]: DEBUG nova.virt.hardware [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.834150] env[61243]: DEBUG nova.virt.hardware [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.834317] env[61243]: DEBUG nova.virt.hardware [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.834561] env[61243]: DEBUG nova.virt.hardware [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.834680] env[61243]: DEBUG nova.virt.hardware [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.834851] env[61243]: DEBUG nova.virt.hardware [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.835802] env[61243]: DEBUG nova.virt.hardware [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.835802] env[61243]: DEBUG nova.virt.hardware [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.835802] env[61243]: DEBUG nova.virt.hardware [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.835802] env[61243]: DEBUG nova.virt.hardware [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.835802] env[61243]: DEBUG nova.virt.hardware [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.836941] env[61243]: DEBUG nova.compute.utils [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 828.840525] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ba7a29-af33-4cc5-8267-edf5f5ddef6b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.843509] env[61243]: DEBUG nova.compute.manager [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Not allocating networking since 'none' was specified. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 828.848577] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8636a83-af4f-4a43-a9fe-46ac30c6afad {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.863456] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Instance VIF info [] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 828.869579] env[61243]: DEBUG oslo.service.loopingcall [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.870225] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 828.870469] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0fe179e7-f426-472a-a060-b279023b0146 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.888900] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 828.888900] env[61243]: value = "task-1338811" [ 828.888900] env[61243]: _type = "Task" [ 828.888900] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.893948] env[61243]: DEBUG nova.compute.manager [req-b6f32886-c523-4b17-bc3b-0a93c6b50f79 req-ad2a4f63-fabf-47f9-a29a-2b65ce7896c6 service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Received event network-vif-plugged-b22c18f9-e3a1-4c8b-a26a-17c651ca16b5 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.894177] env[61243]: DEBUG oslo_concurrency.lockutils [req-b6f32886-c523-4b17-bc3b-0a93c6b50f79 req-ad2a4f63-fabf-47f9-a29a-2b65ce7896c6 service nova] Acquiring lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.894386] env[61243]: DEBUG oslo_concurrency.lockutils [req-b6f32886-c523-4b17-bc3b-0a93c6b50f79 req-ad2a4f63-fabf-47f9-a29a-2b65ce7896c6 service nova] Lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.894552] env[61243]: DEBUG oslo_concurrency.lockutils [req-b6f32886-c523-4b17-bc3b-0a93c6b50f79 req-ad2a4f63-fabf-47f9-a29a-2b65ce7896c6 service nova] Lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.894727] env[61243]: DEBUG nova.compute.manager [req-b6f32886-c523-4b17-bc3b-0a93c6b50f79 req-ad2a4f63-fabf-47f9-a29a-2b65ce7896c6 service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] No waiting events found dispatching network-vif-plugged-b22c18f9-e3a1-4c8b-a26a-17c651ca16b5 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 828.894890] env[61243]: WARNING nova.compute.manager [req-b6f32886-c523-4b17-bc3b-0a93c6b50f79 req-ad2a4f63-fabf-47f9-a29a-2b65ce7896c6 service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Received unexpected event network-vif-plugged-b22c18f9-e3a1-4c8b-a26a-17c651ca16b5 for instance with vm_state building and task_state spawning. [ 828.903728] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338811, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.031613] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Creating linked-clone VM from snapshot {{(pid=61243) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 829.032986] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-680acc17-9a6b-475f-a5bd-530023fac8cf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.046987] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 829.046987] env[61243]: value = "task-1338812" [ 829.046987] env[61243]: _type = "Task" [ 829.046987] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.057565] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338812, 'name': CloneVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.104200] env[61243]: DEBUG nova.network.neutron [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Successfully updated port: b22c18f9-e3a1-4c8b-a26a-17c651ca16b5 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 829.322701] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338810, 'name': ReconfigVM_Task, 'duration_secs': 0.274872} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.323066] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 894eb798-6c9e-47cb-8eb5-4610fe184bc5/894eb798-6c9e-47cb-8eb5-4610fe184bc5.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.324353] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-21ccfc7a-6a26-4954-9588-4b44d2460de0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.330197] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 829.330197] env[61243]: value = "task-1338813" [ 829.330197] env[61243]: _type = "Task" [ 829.330197] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.339210] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338813, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.346486] env[61243]: DEBUG nova.compute.manager [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 829.404048] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338811, 'name': CreateVM_Task, 'duration_secs': 0.280712} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.404048] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 829.405237] env[61243]: DEBUG oslo_vmware.service [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019a6b20-6d8c-4ebe-8984-52fbe74e2fe2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.414480] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.414688] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.415068] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 829.415592] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-364e10d7-2e31-45ba-8f76-d26c98a1a0e5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.423714] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 829.423714] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522526b2-93a7-2ec9-e583-14f4cfb96e91" [ 829.423714] env[61243]: _type = "Task" [ 829.423714] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.434739] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522526b2-93a7-2ec9-e583-14f4cfb96e91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.563507] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338812, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.610284] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "refresh_cache-99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.610284] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquired lock "refresh_cache-99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.610284] env[61243]: DEBUG nova.network.neutron [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 829.677970] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f4de269-0df9-47fb-a515-47dcfaa7d91f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.690241] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0c0bad-ced1-4dab-b342-8ce630f2fa36 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.745535] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf8d628-e5f0-4d5e-96d6-140bd325fba2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.756605] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-739d14b2-0c6f-42cd-bd6d-c1ec018e784a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.778569] env[61243]: DEBUG nova.compute.provider_tree [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.840940] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338813, 'name': Rename_Task, 'duration_secs': 0.134416} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.842104] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 829.842104] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-745a91bd-3a5f-40c0-8a97-c07237da7bd6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.850465] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 829.850465] env[61243]: value = "task-1338814" [ 829.850465] env[61243]: _type = "Task" [ 829.850465] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.863600] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338814, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.933975] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.934372] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 829.934687] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.934873] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.935105] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 829.935424] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-219b37eb-78ff-4ae4-95a9-b37da5f48781 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.944387] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 829.944606] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 829.945558] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d2da9e-3f9d-4ca2-8007-68a420260381 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.952576] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42c5ffca-696a-48e1-ae51-e561d25d1ac8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.958225] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 829.958225] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5265e391-308b-a141-fa85-2be128c47fc4" [ 829.958225] env[61243]: _type = "Task" [ 829.958225] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.969007] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5265e391-308b-a141-fa85-2be128c47fc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.065639] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338812, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.146234] env[61243]: DEBUG nova.network.neutron [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 830.282444] env[61243]: DEBUG nova.scheduler.client.report [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 830.310370] env[61243]: DEBUG nova.network.neutron [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Updating instance_info_cache with network_info: [{"id": "b22c18f9-e3a1-4c8b-a26a-17c651ca16b5", "address": "fa:16:3e:55:c6:a8", "network": {"id": "523d4cf1-467c-4ea6-80a4-ebbd01016efa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-919325376-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7651b565d1647eb948fe07faba32a41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb22c18f9-e3", "ovs_interfaceid": "b22c18f9-e3a1-4c8b-a26a-17c651ca16b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.359438] env[61243]: DEBUG nova.compute.manager [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 830.366179] env[61243]: DEBUG oslo_vmware.api [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338814, 'name': PowerOnVM_Task, 'duration_secs': 0.483534} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.366547] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 830.366978] env[61243]: INFO nova.compute.manager [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Took 6.85 seconds to spawn the instance on the hypervisor. [ 830.367062] env[61243]: DEBUG nova.compute.manager [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 830.367884] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9974ff5-e1e5-471f-8919-514a25d64adc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.398107] env[61243]: DEBUG nova.virt.hardware [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 830.398107] env[61243]: DEBUG nova.virt.hardware [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 830.398107] env[61243]: DEBUG nova.virt.hardware [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.398107] env[61243]: DEBUG nova.virt.hardware [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 830.398312] env[61243]: DEBUG nova.virt.hardware [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.398312] env[61243]: DEBUG nova.virt.hardware [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 830.398312] env[61243]: DEBUG nova.virt.hardware [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 830.399013] env[61243]: DEBUG nova.virt.hardware [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 830.399309] env[61243]: DEBUG nova.virt.hardware [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 830.399865] env[61243]: DEBUG nova.virt.hardware [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 830.400192] env[61243]: DEBUG nova.virt.hardware [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 830.402040] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-582619d1-839b-4aef-8117-1fd48dae6c58 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.410731] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b739d40b-37d3-43d1-829e-fa01cc43824f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.426211] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Instance VIF info [] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.433365] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Creating folder: Project (bcd10de3da214de1bcfc16a54ec68ed0). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.434767] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b896802d-f9da-4cd9-b712-ea558bde5b74 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.447054] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Created folder: Project (bcd10de3da214de1bcfc16a54ec68ed0) in parent group-v285636. [ 830.447054] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Creating folder: Instances. Parent ref: group-v285677. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.447054] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-09247bf9-fe1b-4290-abdc-0abf7245d4a9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.457134] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Created folder: Instances in parent group-v285677. [ 830.457134] env[61243]: DEBUG oslo.service.loopingcall [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.457134] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 830.457134] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b38c511e-2812-448d-becc-9338db9f01e5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.481271] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Preparing fetch location {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 830.481702] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Creating directory with path [datastore1] vmware_temp/e5922f99-d6b6-4b63-8de9-0585d0dd8400/6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.482169] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.482169] env[61243]: value = "task-1338817" [ 830.482169] env[61243]: _type = "Task" [ 830.482169] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.482668] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-16e67fda-d43b-4b0c-ae76-45a6d0c5bff7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.494050] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338817, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.506022] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Created directory with path [datastore1] vmware_temp/e5922f99-d6b6-4b63-8de9-0585d0dd8400/6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.506022] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Fetch image to [datastore1] vmware_temp/e5922f99-d6b6-4b63-8de9-0585d0dd8400/6142e969-c114-4502-aa93-c018fb915a86/tmp-sparse.vmdk {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 830.506022] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Downloading image file data 6142e969-c114-4502-aa93-c018fb915a86 to [datastore1] vmware_temp/e5922f99-d6b6-4b63-8de9-0585d0dd8400/6142e969-c114-4502-aa93-c018fb915a86/tmp-sparse.vmdk on the data store datastore1 {{(pid=61243) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 830.506022] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ed4f72-eda5-4793-a46e-5d17053f3403 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.514400] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72cc2130-c091-43a4-b142-082ce5568f5c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.526270] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52db1c9-608a-4515-ac39-4d8436f5fc86 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.564213] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd09e36-8fb2-47f7-868f-a1af966da4d6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.574509] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338812, 'name': CloneVM_Task} progress is 95%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.575224] env[61243]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-30f9a548-d3cb-40bc-b773-116a021ca499 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.665305] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Downloading image file data 6142e969-c114-4502-aa93-c018fb915a86 to the data store datastore1 {{(pid=61243) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 830.722427] env[61243]: DEBUG oslo_vmware.rw_handles [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/e5922f99-d6b6-4b63-8de9-0585d0dd8400/6142e969-c114-4502-aa93-c018fb915a86/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61243) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 830.792105] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.464s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.792651] env[61243]: DEBUG nova.compute.manager [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 830.799048] env[61243]: DEBUG oslo_concurrency.lockutils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.823s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.799048] env[61243]: INFO nova.compute.claims [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.815704] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Releasing lock "refresh_cache-99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.816239] env[61243]: DEBUG nova.compute.manager [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Instance network_info: |[{"id": "b22c18f9-e3a1-4c8b-a26a-17c651ca16b5", "address": "fa:16:3e:55:c6:a8", "network": {"id": "523d4cf1-467c-4ea6-80a4-ebbd01016efa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-919325376-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7651b565d1647eb948fe07faba32a41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb22c18f9-e3", "ovs_interfaceid": "b22c18f9-e3a1-4c8b-a26a-17c651ca16b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 830.816857] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:c6:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '42f08482-a1da-405d-9918-d733d9f5173c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b22c18f9-e3a1-4c8b-a26a-17c651ca16b5', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.826562] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Creating folder: Project (c7651b565d1647eb948fe07faba32a41). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.827344] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-470eb5d9-c8df-4aa5-9991-d3944f3cde2f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.842081] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Created folder: Project (c7651b565d1647eb948fe07faba32a41) in parent group-v285636. [ 830.842081] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Creating folder: Instances. Parent ref: group-v285680. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.842081] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5c13a095-4082-4d08-8f6c-74f04e401002 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.858449] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Created folder: Instances in parent group-v285680. [ 830.858715] env[61243]: DEBUG oslo.service.loopingcall [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.859555] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 830.859555] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98e56d84-2f84-4cf2-85c9-8823e9cdd159 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.886798] env[61243]: INFO nova.compute.manager [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Took 28.14 seconds to build instance. [ 830.890767] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.890767] env[61243]: value = "task-1338820" [ 830.890767] env[61243]: _type = "Task" [ 830.890767] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.898941] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338820, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.997124] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338817, 'name': CreateVM_Task, 'duration_secs': 0.492142} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.997647] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 830.998240] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.998370] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.998727] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 830.999140] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fb2c041-c8f7-4084-b461-c78459602a4b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.004752] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 831.004752] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d10d58-25a4-4d94-5a31-4f411e1c1aff" [ 831.004752] env[61243]: _type = "Task" [ 831.004752] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.014369] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d10d58-25a4-4d94-5a31-4f411e1c1aff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.062100] env[61243]: DEBUG nova.compute.manager [req-d99cabe4-2f6d-4243-b578-a51e02371055 req-a2f5aa93-071f-456d-88e2-a871bc5cb200 service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Received event network-changed-b22c18f9-e3a1-4c8b-a26a-17c651ca16b5 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.062100] env[61243]: DEBUG nova.compute.manager [req-d99cabe4-2f6d-4243-b578-a51e02371055 req-a2f5aa93-071f-456d-88e2-a871bc5cb200 service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Refreshing instance network info cache due to event network-changed-b22c18f9-e3a1-4c8b-a26a-17c651ca16b5. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 831.063382] env[61243]: DEBUG oslo_concurrency.lockutils [req-d99cabe4-2f6d-4243-b578-a51e02371055 req-a2f5aa93-071f-456d-88e2-a871bc5cb200 service nova] Acquiring lock "refresh_cache-99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.063382] env[61243]: DEBUG oslo_concurrency.lockutils [req-d99cabe4-2f6d-4243-b578-a51e02371055 req-a2f5aa93-071f-456d-88e2-a871bc5cb200 service nova] Acquired lock "refresh_cache-99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.063382] env[61243]: DEBUG nova.network.neutron [req-d99cabe4-2f6d-4243-b578-a51e02371055 req-a2f5aa93-071f-456d-88e2-a871bc5cb200 service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Refreshing network info cache for port b22c18f9-e3a1-4c8b-a26a-17c651ca16b5 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 831.074772] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338812, 'name': CloneVM_Task, 'duration_secs': 1.700184} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.074956] env[61243]: INFO nova.virt.vmwareapi.vmops [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Created linked-clone VM from snapshot [ 831.076780] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d0d0fe-fba6-40a0-a5cf-a97d94219a8b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.086166] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Uploading image 137baf91-2436-47ab-9567-ee1d5ae5b3b2 {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 831.131785] env[61243]: DEBUG oslo_vmware.rw_handles [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 831.131785] env[61243]: value = "vm-285676" [ 831.131785] env[61243]: _type = "VirtualMachine" [ 831.131785] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 831.131785] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-40b28d9e-7cbd-492e-b887-a08cff6b6141 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.138467] env[61243]: DEBUG oslo_vmware.rw_handles [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lease: (returnval){ [ 831.138467] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52384bcd-56fd-3054-7032-b94443fc02d1" [ 831.138467] env[61243]: _type = "HttpNfcLease" [ 831.138467] env[61243]: } obtained for exporting VM: (result){ [ 831.138467] env[61243]: value = "vm-285676" [ 831.138467] env[61243]: _type = "VirtualMachine" [ 831.138467] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 831.138752] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the lease: (returnval){ [ 831.138752] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52384bcd-56fd-3054-7032-b94443fc02d1" [ 831.138752] env[61243]: _type = "HttpNfcLease" [ 831.138752] env[61243]: } to be ready. {{(pid=61243) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 831.146654] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 831.146654] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52384bcd-56fd-3054-7032-b94443fc02d1" [ 831.146654] env[61243]: _type = "HttpNfcLease" [ 831.146654] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 831.297482] env[61243]: DEBUG nova.compute.utils [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 831.299478] env[61243]: DEBUG nova.compute.manager [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Not allocating networking since 'none' was specified. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 831.390140] env[61243]: DEBUG oslo_concurrency.lockutils [None req-53dbaed0-8844-435a-bb6e-98b00430179d tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "894eb798-6c9e-47cb-8eb5-4610fe184bc5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.341s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.412338] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338820, 'name': CreateVM_Task, 'duration_secs': 0.373676} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.413236] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 831.414113] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.415446] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.415446] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 831.415446] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a3c6f8c-8f42-46c5-ab87-53aeaf023f86 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.424146] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 831.424146] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ed017e-8980-8151-c28a-a9f00f549ab1" [ 831.424146] env[61243]: _type = "Task" [ 831.424146] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.445437] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.445437] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 831.445655] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.508019] env[61243]: DEBUG oslo_vmware.rw_handles [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Completed reading data from the image iterator. {{(pid=61243) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 831.508019] env[61243]: DEBUG oslo_vmware.rw_handles [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/e5922f99-d6b6-4b63-8de9-0585d0dd8400/6142e969-c114-4502-aa93-c018fb915a86/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 831.518997] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d10d58-25a4-4d94-5a31-4f411e1c1aff, 'name': SearchDatastore_Task, 'duration_secs': 0.01181} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.519777] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.520174] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 831.520544] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.520817] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.521696] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 831.521696] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ffe2b031-8066-4fe3-bbdc-a8e776e0db61 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.533029] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 831.533029] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 831.533029] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-076fdce2-83ef-4ba3-99cd-b20394d1d20b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.536032] env[61243]: INFO nova.compute.manager [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Rescuing [ 831.541940] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "refresh_cache-894eb798-6c9e-47cb-8eb5-4610fe184bc5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.541940] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquired lock "refresh_cache-894eb798-6c9e-47cb-8eb5-4610fe184bc5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.541940] env[61243]: DEBUG nova.network.neutron [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 831.542496] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 831.542496] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52b5c8b0-9226-90bb-e067-0cf5bad8bff4" [ 831.542496] env[61243]: _type = "Task" [ 831.542496] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.552192] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52b5c8b0-9226-90bb-e067-0cf5bad8bff4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.648486] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 831.648486] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52384bcd-56fd-3054-7032-b94443fc02d1" [ 831.648486] env[61243]: _type = "HttpNfcLease" [ 831.648486] env[61243]: } is ready. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 831.649282] env[61243]: DEBUG oslo_vmware.rw_handles [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 831.649282] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52384bcd-56fd-3054-7032-b94443fc02d1" [ 831.649282] env[61243]: _type = "HttpNfcLease" [ 831.649282] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 831.649564] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a76acb-7a38-4149-881d-2241508901ba {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.652575] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Downloaded image file data 6142e969-c114-4502-aa93-c018fb915a86 to vmware_temp/e5922f99-d6b6-4b63-8de9-0585d0dd8400/6142e969-c114-4502-aa93-c018fb915a86/tmp-sparse.vmdk on the data store datastore1 {{(pid=61243) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 831.656310] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Caching image {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 831.656310] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Copying Virtual Disk [datastore1] vmware_temp/e5922f99-d6b6-4b63-8de9-0585d0dd8400/6142e969-c114-4502-aa93-c018fb915a86/tmp-sparse.vmdk to [datastore1] vmware_temp/e5922f99-d6b6-4b63-8de9-0585d0dd8400/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 831.656310] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fcf20cc7-a901-48d0-bf04-98c7ff6a6a65 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.662933] env[61243]: DEBUG oslo_vmware.rw_handles [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5251b8c9-0d7d-15c6-46d9-0399673adcca/disk-0.vmdk from lease info. {{(pid=61243) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 831.663114] env[61243]: DEBUG oslo_vmware.rw_handles [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5251b8c9-0d7d-15c6-46d9-0399673adcca/disk-0.vmdk for reading. {{(pid=61243) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 831.665612] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 831.665612] env[61243]: value = "task-1338822" [ 831.665612] env[61243]: _type = "Task" [ 831.665612] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.735020] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338822, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.790028] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c907573d-4614-40d7-b56f-7b3bc4c3b5a0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.801280] env[61243]: DEBUG nova.compute.manager [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 831.902684] env[61243]: DEBUG nova.compute.manager [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 831.978689] env[61243]: DEBUG nova.network.neutron [req-d99cabe4-2f6d-4243-b578-a51e02371055 req-a2f5aa93-071f-456d-88e2-a871bc5cb200 service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Updated VIF entry in instance network info cache for port b22c18f9-e3a1-4c8b-a26a-17c651ca16b5. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 831.979053] env[61243]: DEBUG nova.network.neutron [req-d99cabe4-2f6d-4243-b578-a51e02371055 req-a2f5aa93-071f-456d-88e2-a871bc5cb200 service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Updating instance_info_cache with network_info: [{"id": "b22c18f9-e3a1-4c8b-a26a-17c651ca16b5", "address": "fa:16:3e:55:c6:a8", "network": {"id": "523d4cf1-467c-4ea6-80a4-ebbd01016efa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-919325376-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7651b565d1647eb948fe07faba32a41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb22c18f9-e3", "ovs_interfaceid": "b22c18f9-e3a1-4c8b-a26a-17c651ca16b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.053684] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52b5c8b0-9226-90bb-e067-0cf5bad8bff4, 'name': SearchDatastore_Task, 'duration_secs': 0.011363} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.054715] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66dcd889-9716-443e-b0e0-553f62baf9f4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.061211] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 832.061211] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5274e94b-855d-303c-97db-e3ddc1584029" [ 832.061211] env[61243]: _type = "Task" [ 832.061211] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.072599] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5274e94b-855d-303c-97db-e3ddc1584029, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.178028] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338822, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.181895] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e3ff33-a931-41cf-858f-e9e751bed7b2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.191529] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2102cb7e-7f1c-4032-b722-e884e672766c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.228637] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b541574-b229-4824-af0a-1a2160e0b49c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.236947] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337e4dca-6ebf-4ab5-a1fd-50f170d8d3cf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.252890] env[61243]: DEBUG nova.compute.provider_tree [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.430846] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.481979] env[61243]: DEBUG oslo_concurrency.lockutils [req-d99cabe4-2f6d-4243-b578-a51e02371055 req-a2f5aa93-071f-456d-88e2-a871bc5cb200 service nova] Releasing lock "refresh_cache-99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.505587] env[61243]: DEBUG nova.network.neutron [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Updating instance_info_cache with network_info: [{"id": "bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0", "address": "fa:16:3e:f6:c2:39", "network": {"id": "8e2b6835-484b-4346-8bd1-034c53671110", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710127686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4fcef2317fdf47f8ac820aeef1765fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdfbe0d5-9b", "ovs_interfaceid": "bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.574354] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5274e94b-855d-303c-97db-e3ddc1584029, 'name': SearchDatastore_Task, 'duration_secs': 0.015123} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.574447] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.574689] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 729d9bd9-1ab0-47bd-9b7a-11412d3dc608/729d9bd9-1ab0-47bd-9b7a-11412d3dc608.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 832.575642] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5a55fddb-fdaa-4a2a-8e2b-9108c5ebc6d6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.581905] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 832.581905] env[61243]: value = "task-1338823" [ 832.581905] env[61243]: _type = "Task" [ 832.581905] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.592364] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338823, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.680463] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338822, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.758801] env[61243]: DEBUG nova.scheduler.client.report [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.813481] env[61243]: DEBUG nova.compute.manager [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 832.849508] env[61243]: DEBUG nova.virt.hardware [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 832.849920] env[61243]: DEBUG nova.virt.hardware [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 832.850130] env[61243]: DEBUG nova.virt.hardware [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.850672] env[61243]: DEBUG nova.virt.hardware [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 832.850953] env[61243]: DEBUG nova.virt.hardware [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.851183] env[61243]: DEBUG nova.virt.hardware [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 832.851558] env[61243]: DEBUG nova.virt.hardware [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 832.851968] env[61243]: DEBUG nova.virt.hardware [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 832.852260] env[61243]: DEBUG nova.virt.hardware [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 832.852485] env[61243]: DEBUG nova.virt.hardware [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 832.852687] env[61243]: DEBUG nova.virt.hardware [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.854081] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-223e3612-896a-49bc-8850-2ba4b4c66a03 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.865168] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bada0ef-9e63-49f8-8f4d-341168bd235a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.885402] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Instance VIF info [] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 832.892040] env[61243]: DEBUG oslo.service.loopingcall [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.892859] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 832.895232] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eb526327-c5cd-47aa-a6c1-39cdf01152bd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.913251] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 832.913251] env[61243]: value = "task-1338824" [ 832.913251] env[61243]: _type = "Task" [ 832.913251] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.925102] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338824, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.011209] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Releasing lock "refresh_cache-894eb798-6c9e-47cb-8eb5-4610fe184bc5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.096240] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338823, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.179756] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338822, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.394025} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.180231] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Copied Virtual Disk [datastore1] vmware_temp/e5922f99-d6b6-4b63-8de9-0585d0dd8400/6142e969-c114-4502-aa93-c018fb915a86/tmp-sparse.vmdk to [datastore1] vmware_temp/e5922f99-d6b6-4b63-8de9-0585d0dd8400/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 833.180321] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Deleting the datastore file [datastore1] vmware_temp/e5922f99-d6b6-4b63-8de9-0585d0dd8400/6142e969-c114-4502-aa93-c018fb915a86/tmp-sparse.vmdk {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 833.180600] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5e53525-7d5b-47c9-a667-d735f8ba9945 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.190020] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 833.190020] env[61243]: value = "task-1338825" [ 833.190020] env[61243]: _type = "Task" [ 833.190020] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.201166] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338825, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.264646] env[61243]: DEBUG oslo_concurrency.lockutils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.468s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.264646] env[61243]: DEBUG nova.compute.manager [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 833.269271] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.781s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.270905] env[61243]: INFO nova.compute.claims [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 833.424924] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338824, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.561399] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 833.561753] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3252608a-f590-4936-8f4b-db5e5b9a6033 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.570126] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 833.570126] env[61243]: value = "task-1338826" [ 833.570126] env[61243]: _type = "Task" [ 833.570126] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.580802] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338826, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.592555] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338823, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.738264} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.592868] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 729d9bd9-1ab0-47bd-9b7a-11412d3dc608/729d9bd9-1ab0-47bd-9b7a-11412d3dc608.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 833.593436] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 833.593780] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-abab15c8-7f60-4d82-94f1-74aca743cd1b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.603021] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 833.603021] env[61243]: value = "task-1338827" [ 833.603021] env[61243]: _type = "Task" [ 833.603021] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.611186] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338827, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.702429] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338825, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.045223} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.702429] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 833.702609] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Moving file from [datastore1] vmware_temp/e5922f99-d6b6-4b63-8de9-0585d0dd8400/6142e969-c114-4502-aa93-c018fb915a86 to [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86. {{(pid=61243) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 833.702868] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-3ca91521-0f85-425d-82cc-fe545b283907 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.710106] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 833.710106] env[61243]: value = "task-1338828" [ 833.710106] env[61243]: _type = "Task" [ 833.710106] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.721716] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338828, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.775400] env[61243]: DEBUG nova.compute.utils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 833.780220] env[61243]: DEBUG nova.compute.manager [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 833.780220] env[61243]: DEBUG nova.network.neutron [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 833.840744] env[61243]: DEBUG nova.policy [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d149744716c4bf0b0ba57838458f52e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a0e29b6651fd46fcab37859fed86d898', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 833.925400] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338824, 'name': CreateVM_Task, 'duration_secs': 0.569141} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.925732] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 833.926481] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.926823] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.927447] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 833.927895] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-412b6da1-e227-4604-8b09-9878da9d842c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.935499] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 833.935499] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cb4763-ee96-b3e7-263b-a41338647f10" [ 833.935499] env[61243]: _type = "Task" [ 833.935499] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.948629] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cb4763-ee96-b3e7-263b-a41338647f10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.084521] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338826, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.111540] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338827, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069297} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.111852] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 834.116502] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca23c3a-ffbf-4467-87c4-e75d3de0875d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.140195] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 729d9bd9-1ab0-47bd-9b7a-11412d3dc608/729d9bd9-1ab0-47bd-9b7a-11412d3dc608.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.140195] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be0581f0-5b5b-467b-8e58-5c117c6910fa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.163249] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 834.163249] env[61243]: value = "task-1338829" [ 834.163249] env[61243]: _type = "Task" [ 834.163249] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.173177] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338829, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.184381] env[61243]: DEBUG nova.network.neutron [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Successfully created port: f2472adb-26ae-4651-9108-d8286beeb98b {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 834.220628] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338828, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.034801} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.221047] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] File moved {{(pid=61243) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 834.221443] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Cleaning up location [datastore1] vmware_temp/e5922f99-d6b6-4b63-8de9-0585d0dd8400 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 834.221549] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Deleting the datastore file [datastore1] vmware_temp/e5922f99-d6b6-4b63-8de9-0585d0dd8400 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 834.221911] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1056596c-587a-4df8-a30e-95d9e7706851 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.228302] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 834.228302] env[61243]: value = "task-1338830" [ 834.228302] env[61243]: _type = "Task" [ 834.228302] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.236832] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338830, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.280900] env[61243]: DEBUG nova.compute.manager [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 834.448643] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cb4763-ee96-b3e7-263b-a41338647f10, 'name': SearchDatastore_Task, 'duration_secs': 0.031345} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.448643] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.448643] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.448788] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.448889] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.449086] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.449366] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fbd2fb88-fdb7-4c73-b406-5166122df1a5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.459878] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.460054] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 834.460940] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d47fd96-4c5b-41c2-a645-c78720f8e69d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.466363] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 834.466363] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524171df-adc5-419d-62fa-231a02cbd50d" [ 834.466363] env[61243]: _type = "Task" [ 834.466363] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.478945] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524171df-adc5-419d-62fa-231a02cbd50d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.582429] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338826, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.642097] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Acquiring lock "8df486c4-d1ce-47b0-81a8-ad9e83eaac7f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.642434] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Lock "8df486c4-d1ce-47b0-81a8-ad9e83eaac7f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.648729] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698d5d83-62a1-44d4-8d3e-5f25dd023878 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.659034] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745733c1-fb35-4e8b-b750-784c3cf3cfb5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.671544] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338829, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.696394] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13be4b8a-4022-4727-b84e-5fe1304d9bad {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.703743] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d6d757-0dca-4509-bb77-1a427e60180a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.716862] env[61243]: DEBUG nova.compute.provider_tree [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.738712] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338830, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.041052} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.738712] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.738712] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58d1857c-5bf7-4ed8-8736-4d8e59210480 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.743416] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 834.743416] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d3f7b2-03e5-67a5-28d1-c41ffe3ddedf" [ 834.743416] env[61243]: _type = "Task" [ 834.743416] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.752420] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d3f7b2-03e5-67a5-28d1-c41ffe3ddedf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.977823] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524171df-adc5-419d-62fa-231a02cbd50d, 'name': SearchDatastore_Task, 'duration_secs': 0.010755} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.978731] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5069b728-887c-48d5-81a6-07e70300a46b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.984633] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 834.984633] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5234fde2-e3cc-9fac-a366-b921f2ccb358" [ 834.984633] env[61243]: _type = "Task" [ 834.984633] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.992976] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5234fde2-e3cc-9fac-a366-b921f2ccb358, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.082889] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338826, 'name': PowerOffVM_Task, 'duration_secs': 1.103956} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.083256] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 835.084119] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c97c062a-855f-4072-add7-b2929a953aac {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.103566] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a30850-fc5e-4558-aca1-1547eddb34b0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.131663] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 835.131985] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-447db29c-ad4c-47b2-9dac-fa4528714aa4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.139424] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 835.139424] env[61243]: value = "task-1338831" [ 835.139424] env[61243]: _type = "Task" [ 835.139424] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.147661] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338831, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.172288] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338829, 'name': ReconfigVM_Task, 'duration_secs': 0.63446} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.172578] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 729d9bd9-1ab0-47bd-9b7a-11412d3dc608/729d9bd9-1ab0-47bd-9b7a-11412d3dc608.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.173236] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa29b45b-a291-46c3-86f1-106dd3cc54b9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.179961] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 835.179961] env[61243]: value = "task-1338832" [ 835.179961] env[61243]: _type = "Task" [ 835.179961] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.189886] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338832, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.220188] env[61243]: DEBUG nova.scheduler.client.report [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 835.255279] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d3f7b2-03e5-67a5-28d1-c41ffe3ddedf, 'name': SearchDatastore_Task, 'duration_secs': 0.010939} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.255279] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.255279] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 6cee3dd1-c13d-427d-889c-6cd3ae388a31/6cee3dd1-c13d-427d-889c-6cd3ae388a31.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 835.255279] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.255472] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.255570] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-74eaf0c6-b985-4132-9282-f8d984c2f046 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.257646] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-22d849e5-a297-41e4-9619-d880868c8787 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.264709] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 835.264709] env[61243]: value = "task-1338833" [ 835.264709] env[61243]: _type = "Task" [ 835.264709] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.269041] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.269041] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 835.270034] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c51ca0fc-06f6-4ec4-9786-e7974132f4e1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.275646] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338833, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.281374] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 835.281374] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522099b6-4e43-7faa-0ff5-d45b7cd510a9" [ 835.281374] env[61243]: _type = "Task" [ 835.281374] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.287586] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522099b6-4e43-7faa-0ff5-d45b7cd510a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.293971] env[61243]: DEBUG nova.compute.manager [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 835.326718] env[61243]: DEBUG nova.virt.hardware [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.327066] env[61243]: DEBUG nova.virt.hardware [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.328505] env[61243]: DEBUG nova.virt.hardware [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.328505] env[61243]: DEBUG nova.virt.hardware [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.328505] env[61243]: DEBUG nova.virt.hardware [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.328505] env[61243]: DEBUG nova.virt.hardware [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.328505] env[61243]: DEBUG nova.virt.hardware [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.328732] env[61243]: DEBUG nova.virt.hardware [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.328732] env[61243]: DEBUG nova.virt.hardware [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.328732] env[61243]: DEBUG nova.virt.hardware [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.328827] env[61243]: DEBUG nova.virt.hardware [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.329724] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b228d551-6bda-4cd9-8191-332a9e30a012 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.338287] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed2d3d9-d303-4847-8f8b-5c543e70f1af {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.499250] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5234fde2-e3cc-9fac-a366-b921f2ccb358, 'name': SearchDatastore_Task, 'duration_secs': 0.01216} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.499628] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.499907] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 96315d9e-4eda-4e3a-af0d-bdc52ab181e3/96315d9e-4eda-4e3a-af0d-bdc52ab181e3.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 835.500887] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e185048d-1771-4a53-877b-98413cc6373a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.508706] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 835.508706] env[61243]: value = "task-1338834" [ 835.508706] env[61243]: _type = "Task" [ 835.508706] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.521675] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338834, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.650455] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] VM already powered off {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 835.650718] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.651059] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.651257] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.651491] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.651800] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-58c80a4b-3bd7-445b-b3a1-1182681508a5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.662965] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.663107] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 835.664029] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08873baa-f8b2-4ca1-bb8b-128b3045aa1c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.671677] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 835.671677] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529eff75-9779-f1af-ddb1-c5ed8143d157" [ 835.671677] env[61243]: _type = "Task" [ 835.671677] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.680828] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529eff75-9779-f1af-ddb1-c5ed8143d157, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.689151] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338832, 'name': Rename_Task, 'duration_secs': 0.154459} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.690675] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 835.691852] env[61243]: DEBUG nova.compute.manager [req-b7346794-b45a-4f0b-b966-9e2a6e17f11c req-428b6e2e-36f1-4db3-8291-1cd9ab91693b service nova] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Received event network-vif-plugged-f2472adb-26ae-4651-9108-d8286beeb98b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.692074] env[61243]: DEBUG oslo_concurrency.lockutils [req-b7346794-b45a-4f0b-b966-9e2a6e17f11c req-428b6e2e-36f1-4db3-8291-1cd9ab91693b service nova] Acquiring lock "6f33b303-1ecd-478e-8fa3-33a5f9a5af50-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.692292] env[61243]: DEBUG oslo_concurrency.lockutils [req-b7346794-b45a-4f0b-b966-9e2a6e17f11c req-428b6e2e-36f1-4db3-8291-1cd9ab91693b service nova] Lock "6f33b303-1ecd-478e-8fa3-33a5f9a5af50-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.692482] env[61243]: DEBUG oslo_concurrency.lockutils [req-b7346794-b45a-4f0b-b966-9e2a6e17f11c req-428b6e2e-36f1-4db3-8291-1cd9ab91693b service nova] Lock "6f33b303-1ecd-478e-8fa3-33a5f9a5af50-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.692673] env[61243]: DEBUG nova.compute.manager [req-b7346794-b45a-4f0b-b966-9e2a6e17f11c req-428b6e2e-36f1-4db3-8291-1cd9ab91693b service nova] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] No waiting events found dispatching network-vif-plugged-f2472adb-26ae-4651-9108-d8286beeb98b {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 835.692852] env[61243]: WARNING nova.compute.manager [req-b7346794-b45a-4f0b-b966-9e2a6e17f11c req-428b6e2e-36f1-4db3-8291-1cd9ab91693b service nova] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Received unexpected event network-vif-plugged-f2472adb-26ae-4651-9108-d8286beeb98b for instance with vm_state building and task_state spawning. [ 835.693190] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7663d2cb-6842-43fb-bcf1-2604ca0fd8f4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.700486] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 835.700486] env[61243]: value = "task-1338835" [ 835.700486] env[61243]: _type = "Task" [ 835.700486] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.710282] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338835, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.725020] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.457s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.725662] env[61243]: DEBUG nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 835.728792] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.759s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.730736] env[61243]: INFO nova.compute.claims [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 835.775866] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338833, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.798573] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522099b6-4e43-7faa-0ff5-d45b7cd510a9, 'name': SearchDatastore_Task, 'duration_secs': 0.014894} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.798573] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-928e025c-7dca-427f-b7b2-c6500a3e2744 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.813049] env[61243]: DEBUG nova.network.neutron [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Successfully updated port: f2472adb-26ae-4651-9108-d8286beeb98b {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 835.813049] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 835.813049] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e09811-52da-8b66-4a79-e62b711df0ec" [ 835.813049] env[61243]: _type = "Task" [ 835.813049] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.824299] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e09811-52da-8b66-4a79-e62b711df0ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.021035] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338834, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.077030] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "3f839aa5-2b9a-4807-b63b-931f74455532" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.077180] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "3f839aa5-2b9a-4807-b63b-931f74455532" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.182776] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529eff75-9779-f1af-ddb1-c5ed8143d157, 'name': SearchDatastore_Task, 'duration_secs': 0.013349} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.183914] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7a77d5d-4d5d-4c25-8d95-c8eebfef8fc0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.188976] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 836.188976] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5237b9af-438d-3278-6f9a-00e8015aa4a1" [ 836.188976] env[61243]: _type = "Task" [ 836.188976] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.196695] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5237b9af-438d-3278-6f9a-00e8015aa4a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.209198] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338835, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.235898] env[61243]: DEBUG nova.compute.utils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 836.240382] env[61243]: DEBUG nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 836.240382] env[61243]: DEBUG nova.network.neutron [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 836.275752] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338833, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.714928} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.276117] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 6cee3dd1-c13d-427d-889c-6cd3ae388a31/6cee3dd1-c13d-427d-889c-6cd3ae388a31.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 836.276361] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 836.276838] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6adcdc55-0878-4cae-a084-94a9bbcabb72 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.284018] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 836.284018] env[61243]: value = "task-1338836" [ 836.284018] env[61243]: _type = "Task" [ 836.284018] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.294293] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338836, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.295921] env[61243]: DEBUG nova.policy [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18fd39290a104706a2e2633308d41213', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '39a655fc062e4322a5a129b66a45684f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 836.314468] env[61243]: DEBUG oslo_concurrency.lockutils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Acquiring lock "refresh_cache-6f33b303-1ecd-478e-8fa3-33a5f9a5af50" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.314635] env[61243]: DEBUG oslo_concurrency.lockutils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Acquired lock "refresh_cache-6f33b303-1ecd-478e-8fa3-33a5f9a5af50" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.316177] env[61243]: DEBUG nova.network.neutron [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 836.328881] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e09811-52da-8b66-4a79-e62b711df0ec, 'name': SearchDatastore_Task, 'duration_secs': 0.087756} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.330008] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.330377] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba/99ce6ca3-a478-4ebe-bf1b-42459aacd9ba.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 836.330686] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fb0dcc19-63d0-4326-8d61-553b6535708c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.339140] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 836.339140] env[61243]: value = "task-1338837" [ 836.339140] env[61243]: _type = "Task" [ 836.339140] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.348852] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1338837, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.521238] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338834, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.659536} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.521558] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 96315d9e-4eda-4e3a-af0d-bdc52ab181e3/96315d9e-4eda-4e3a-af0d-bdc52ab181e3.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 836.521790] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 836.522074] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5e388351-998f-4606-b7e8-7aa7eea7cc80 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.529902] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 836.529902] env[61243]: value = "task-1338838" [ 836.529902] env[61243]: _type = "Task" [ 836.529902] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.539335] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338838, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.588130] env[61243]: DEBUG nova.network.neutron [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Successfully created port: f80550c9-ffa4-4f9f-a0c1-7be2562defa3 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 836.699727] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5237b9af-438d-3278-6f9a-00e8015aa4a1, 'name': SearchDatastore_Task, 'duration_secs': 0.009444} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.700081] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.700395] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 894eb798-6c9e-47cb-8eb5-4610fe184bc5/6142e969-c114-4502-aa93-c018fb915a86-rescue.vmdk. {{(pid=61243) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 836.700766] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dfd6a18d-01cb-4a2c-9719-0cc7c7b33e77 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.713099] env[61243]: DEBUG oslo_vmware.api [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338835, 'name': PowerOnVM_Task, 'duration_secs': 0.644196} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.714681] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 836.714917] env[61243]: INFO nova.compute.manager [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Took 6.36 seconds to spawn the instance on the hypervisor. [ 836.715242] env[61243]: DEBUG nova.compute.manager [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 836.715540] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 836.715540] env[61243]: value = "task-1338839" [ 836.715540] env[61243]: _type = "Task" [ 836.715540] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.719512] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea7092f7-46b0-4628-8a43-5aade13c4f8b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.734114] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338839, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.740374] env[61243]: DEBUG nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 836.799218] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338836, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080834} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.800100] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.800445] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-336b111c-7d79-4b3b-8241-960a4ccc4c8f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.826929] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 6cee3dd1-c13d-427d-889c-6cd3ae388a31/6cee3dd1-c13d-427d-889c-6cd3ae388a31.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.831850] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-036ba925-5c6a-471c-ac83-0c6ac0c1565b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.857332] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1338837, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.859047] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 836.859047] env[61243]: value = "task-1338840" [ 836.859047] env[61243]: _type = "Task" [ 836.859047] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.870891] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338840, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.890113] env[61243]: DEBUG nova.network.neutron [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 837.044029] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338838, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100359} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.044029] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 837.044029] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b6214e-bdf1-4cc0-9001-77fee31d210b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.048619] env[61243]: DEBUG nova.network.neutron [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Updating instance_info_cache with network_info: [{"id": "f2472adb-26ae-4651-9108-d8286beeb98b", "address": "fa:16:3e:c7:ef:b5", "network": {"id": "8dddc1c6-47b7-416e-a921-1431851b15db", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1168224682-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a0e29b6651fd46fcab37859fed86d898", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2472adb-26", "ovs_interfaceid": "f2472adb-26ae-4651-9108-d8286beeb98b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.068195] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 96315d9e-4eda-4e3a-af0d-bdc52ab181e3/96315d9e-4eda-4e3a-af0d-bdc52ab181e3.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.071906] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05d1080f-512b-4601-94da-2b365ee7bab7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.092292] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 837.092292] env[61243]: value = "task-1338841" [ 837.092292] env[61243]: _type = "Task" [ 837.092292] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.105165] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338841, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.136546] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa03a4c2-d40d-4bee-9ee1-54d2a84636cd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.144780] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc2523e-1b8f-4d03-883d-99d9ed6a08b1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.177511] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27bc7869-4943-4b20-8c4d-5ada6a600cee {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.186419] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c5c363-8c2e-4fff-ad72-85b44e01e017 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.201623] env[61243]: DEBUG nova.compute.provider_tree [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.229124] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338839, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.241758] env[61243]: INFO nova.compute.manager [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Took 19.93 seconds to build instance. [ 837.359558] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1338837, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.560516} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.359941] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba/99ce6ca3-a478-4ebe-bf1b-42459aacd9ba.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 837.360175] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.360480] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-85bf0220-f88c-4eb2-8855-a9b52b0c360b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.371669] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338840, 'name': ReconfigVM_Task, 'duration_secs': 0.338973} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.373134] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 6cee3dd1-c13d-427d-889c-6cd3ae388a31/6cee3dd1-c13d-427d-889c-6cd3ae388a31.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.373850] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 837.373850] env[61243]: value = "task-1338842" [ 837.373850] env[61243]: _type = "Task" [ 837.373850] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.374174] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e1476e4b-ac35-4998-b782-974286ce0fab {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.383664] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1338842, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.385087] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 837.385087] env[61243]: value = "task-1338843" [ 837.385087] env[61243]: _type = "Task" [ 837.385087] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.393550] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338843, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.552193] env[61243]: DEBUG oslo_concurrency.lockutils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Releasing lock "refresh_cache-6f33b303-1ecd-478e-8fa3-33a5f9a5af50" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.552593] env[61243]: DEBUG nova.compute.manager [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Instance network_info: |[{"id": "f2472adb-26ae-4651-9108-d8286beeb98b", "address": "fa:16:3e:c7:ef:b5", "network": {"id": "8dddc1c6-47b7-416e-a921-1431851b15db", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1168224682-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a0e29b6651fd46fcab37859fed86d898", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2472adb-26", "ovs_interfaceid": "f2472adb-26ae-4651-9108-d8286beeb98b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 837.553488] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c7:ef:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98e21102-8954-4f6f-b1e6-5d764a53aa22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f2472adb-26ae-4651-9108-d8286beeb98b', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.562857] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Creating folder: Project (a0e29b6651fd46fcab37859fed86d898). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.563314] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e9808ce0-92dc-43d8-9ccd-704ade32af81 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.574053] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Created folder: Project (a0e29b6651fd46fcab37859fed86d898) in parent group-v285636. [ 837.574304] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Creating folder: Instances. Parent ref: group-v285684. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.574599] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8337cb9b-001d-4bf0-8380-10fe62dcd6ae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.584092] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Created folder: Instances in parent group-v285684. [ 837.584368] env[61243]: DEBUG oslo.service.loopingcall [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.584575] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 837.584791] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-443cd893-5d82-4b4c-ab34-a4f9258b8033 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.607540] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.608764] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.608764] env[61243]: value = "task-1338846" [ 837.608764] env[61243]: _type = "Task" [ 837.608764] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.616320] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338846, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.706492] env[61243]: DEBUG nova.scheduler.client.report [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 837.732232] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338839, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.645593} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.733630] env[61243]: INFO nova.virt.vmwareapi.ds_util [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 894eb798-6c9e-47cb-8eb5-4610fe184bc5/6142e969-c114-4502-aa93-c018fb915a86-rescue.vmdk. [ 837.734816] env[61243]: DEBUG nova.compute.manager [req-c350ba71-5ac8-4c00-b276-abb8751fd5b4 req-c98a62d6-ffa0-4437-b186-e58a412f63ec service nova] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Received event network-changed-f2472adb-26ae-4651-9108-d8286beeb98b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.735015] env[61243]: DEBUG nova.compute.manager [req-c350ba71-5ac8-4c00-b276-abb8751fd5b4 req-c98a62d6-ffa0-4437-b186-e58a412f63ec service nova] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Refreshing instance network info cache due to event network-changed-f2472adb-26ae-4651-9108-d8286beeb98b. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 837.735239] env[61243]: DEBUG oslo_concurrency.lockutils [req-c350ba71-5ac8-4c00-b276-abb8751fd5b4 req-c98a62d6-ffa0-4437-b186-e58a412f63ec service nova] Acquiring lock "refresh_cache-6f33b303-1ecd-478e-8fa3-33a5f9a5af50" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.735386] env[61243]: DEBUG oslo_concurrency.lockutils [req-c350ba71-5ac8-4c00-b276-abb8751fd5b4 req-c98a62d6-ffa0-4437-b186-e58a412f63ec service nova] Acquired lock "refresh_cache-6f33b303-1ecd-478e-8fa3-33a5f9a5af50" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.735551] env[61243]: DEBUG nova.network.neutron [req-c350ba71-5ac8-4c00-b276-abb8751fd5b4 req-c98a62d6-ffa0-4437-b186-e58a412f63ec service nova] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Refreshing network info cache for port f2472adb-26ae-4651-9108-d8286beeb98b {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 837.737156] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ecb4ead-256c-4740-bf00-0353e4650e71 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.755989] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1405454e-2e7f-4c1c-98d0-e5dd782d03a1 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "729d9bd9-1ab0-47bd-9b7a-11412d3dc608" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.473s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.757816] env[61243]: DEBUG nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 837.767472] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 894eb798-6c9e-47cb-8eb5-4610fe184bc5/6142e969-c114-4502-aa93-c018fb915a86-rescue.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.769128] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e3bf919-dd00-4bd6-aad8-e27731331f84 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.785043] env[61243]: DEBUG nova.compute.manager [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 837.794971] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 837.795495] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 837.795495] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 837.795612] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 837.795738] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 837.795891] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 837.796107] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 837.796272] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 837.796449] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 837.796668] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 837.797104] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 837.797705] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 837.797705] env[61243]: value = "task-1338847" [ 837.797705] env[61243]: _type = "Task" [ 837.797705] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.798318] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbfefb83-1b99-47b5-b544-c2ecd2c0786a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.813195] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ae9f6a-88af-47f7-bcba-89648f4362c6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.818734] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.885381] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1338842, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.200774} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.885854] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 837.889498] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d36c99b-4f0e-42d8-8ec6-42fdf7055be4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.912012] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba/99ce6ca3-a478-4ebe-bf1b-42459aacd9ba.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.914839] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce2feb5d-a4f6-4789-ac37-4960cba193e3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.928800] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338843, 'name': Rename_Task, 'duration_secs': 0.229597} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.929108] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 837.929806] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4be0d14-090b-4bd7-9e78-656fbcb18abc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.933872] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 837.933872] env[61243]: value = "task-1338848" [ 837.933872] env[61243]: _type = "Task" [ 837.933872] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.938081] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 837.938081] env[61243]: value = "task-1338849" [ 837.938081] env[61243]: _type = "Task" [ 837.938081] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.946578] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1338848, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.952276] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338849, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.111816] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.120846] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338846, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.182275] env[61243]: DEBUG nova.network.neutron [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Successfully updated port: f80550c9-ffa4-4f9f-a0c1-7be2562defa3 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 838.213218] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.213808] env[61243]: DEBUG nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 838.216397] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.712s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.218277] env[61243]: INFO nova.compute.claims [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.313598] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.355358] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.447536] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1338848, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.451355] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338849, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.494744] env[61243]: DEBUG nova.network.neutron [req-c350ba71-5ac8-4c00-b276-abb8751fd5b4 req-c98a62d6-ffa0-4437-b186-e58a412f63ec service nova] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Updated VIF entry in instance network info cache for port f2472adb-26ae-4651-9108-d8286beeb98b. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 838.495127] env[61243]: DEBUG nova.network.neutron [req-c350ba71-5ac8-4c00-b276-abb8751fd5b4 req-c98a62d6-ffa0-4437-b186-e58a412f63ec service nova] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Updating instance_info_cache with network_info: [{"id": "f2472adb-26ae-4651-9108-d8286beeb98b", "address": "fa:16:3e:c7:ef:b5", "network": {"id": "8dddc1c6-47b7-416e-a921-1431851b15db", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1168224682-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a0e29b6651fd46fcab37859fed86d898", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2472adb-26", "ovs_interfaceid": "f2472adb-26ae-4651-9108-d8286beeb98b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.611067] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338841, 'name': ReconfigVM_Task, 'duration_secs': 1.24422} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.611067] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 96315d9e-4eda-4e3a-af0d-bdc52ab181e3/96315d9e-4eda-4e3a-af0d-bdc52ab181e3.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.611067] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-154787a8-d2be-4de9-b7b8-8220494993f8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.621083] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338846, 'name': CreateVM_Task, 'duration_secs': 0.61809} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.622052] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 838.622404] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 838.622404] env[61243]: value = "task-1338850" [ 838.622404] env[61243]: _type = "Task" [ 838.622404] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.623009] env[61243]: DEBUG oslo_concurrency.lockutils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.623187] env[61243]: DEBUG oslo_concurrency.lockutils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.623512] env[61243]: DEBUG oslo_concurrency.lockutils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 838.623787] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fd41bd3-3401-4466-8447-178ffa9ff048 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.629954] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Waiting for the task: (returnval){ [ 838.629954] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52953ef2-b9c5-efbb-cb8a-b8a1cfa3c096" [ 838.629954] env[61243]: _type = "Task" [ 838.629954] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.632911] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338850, 'name': Rename_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.640167] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52953ef2-b9c5-efbb-cb8a-b8a1cfa3c096, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.684978] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "refresh_cache-9a66fd05-1880-423a-8f60-4f11f1c75ab3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.685161] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquired lock "refresh_cache-9a66fd05-1880-423a-8f60-4f11f1c75ab3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.685319] env[61243]: DEBUG nova.network.neutron [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 838.725586] env[61243]: DEBUG nova.compute.utils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 838.726983] env[61243]: DEBUG nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 838.726983] env[61243]: DEBUG nova.network.neutron [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 838.765764] env[61243]: DEBUG nova.policy [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18fd39290a104706a2e2633308d41213', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '39a655fc062e4322a5a129b66a45684f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 838.813622] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338847, 'name': ReconfigVM_Task, 'duration_secs': 0.647581} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.813622] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 894eb798-6c9e-47cb-8eb5-4610fe184bc5/6142e969-c114-4502-aa93-c018fb915a86-rescue.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.814754] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79db0247-d696-4568-92f5-5ad780b8be41 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.847562] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82b7ca67-61ae-4e67-9ef7-c48e30f7f376 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.864068] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 838.864068] env[61243]: value = "task-1338851" [ 838.864068] env[61243]: _type = "Task" [ 838.864068] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.873667] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338851, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.960817] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1338848, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.961300] env[61243]: DEBUG oslo_vmware.api [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338849, 'name': PowerOnVM_Task, 'duration_secs': 0.65188} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.961617] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 838.961856] env[61243]: DEBUG nova.compute.manager [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 838.962684] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320adb4d-41bd-4660-8e78-9cc2105f2079 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.997886] env[61243]: DEBUG oslo_concurrency.lockutils [req-c350ba71-5ac8-4c00-b276-abb8751fd5b4 req-c98a62d6-ffa0-4437-b186-e58a412f63ec service nova] Releasing lock "refresh_cache-6f33b303-1ecd-478e-8fa3-33a5f9a5af50" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.134389] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338850, 'name': Rename_Task, 'duration_secs': 0.249532} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.137447] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 839.137765] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ae5dfea-f8f6-448f-a332-b6a5be0a76e0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.146060] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52953ef2-b9c5-efbb-cb8a-b8a1cfa3c096, 'name': SearchDatastore_Task, 'duration_secs': 0.011433} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.147531] env[61243]: DEBUG oslo_concurrency.lockutils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.147826] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.148117] env[61243]: DEBUG oslo_concurrency.lockutils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.148286] env[61243]: DEBUG oslo_concurrency.lockutils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.148496] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.148918] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 839.148918] env[61243]: value = "task-1338852" [ 839.148918] env[61243]: _type = "Task" [ 839.148918] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.149194] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f6bdcc75-ffa6-4ae1-8433-ee8ad5b83054 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.161132] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338852, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.162743] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.162743] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 839.163970] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60c3fac2-c4ea-4a17-a258-5627f1737593 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.169365] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Waiting for the task: (returnval){ [ 839.169365] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5240e65a-1964-44ba-edeb-2c12681a786f" [ 839.169365] env[61243]: _type = "Task" [ 839.169365] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.178289] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5240e65a-1964-44ba-edeb-2c12681a786f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.233566] env[61243]: DEBUG nova.network.neutron [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 839.233566] env[61243]: DEBUG nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 839.266837] env[61243]: DEBUG nova.network.neutron [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Successfully created port: ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 839.373377] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338851, 'name': ReconfigVM_Task, 'duration_secs': 0.195909} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.373667] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 839.373927] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a1a2356-b108-4597-9d85-2b44a9e45504 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.380397] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 839.380397] env[61243]: value = "task-1338853" [ 839.380397] env[61243]: _type = "Task" [ 839.380397] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.388712] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338853, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.453021] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1338848, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.485921] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.498128] env[61243]: DEBUG nova.network.neutron [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Updating instance_info_cache with network_info: [{"id": "f80550c9-ffa4-4f9f-a0c1-7be2562defa3", "address": "fa:16:3e:83:38:f1", "network": {"id": "e10de432-a279-478d-a226-8cb1663f8f1e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-237966643-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39a655fc062e4322a5a129b66a45684f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf80550c9-ff", "ovs_interfaceid": "f80550c9-ffa4-4f9f-a0c1-7be2562defa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.568017] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba34e680-eaa1-40b8-89c6-bbc6a0ed00ff {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.577229] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded8a14a-ab53-48f4-a09c-ecf8c990898f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.612870] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f0320a-bc0e-4c47-affd-6e82a181daa6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.617816] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b35811-0b52-4fb9-b21e-a89a5c39da00 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.632111] env[61243]: DEBUG nova.compute.provider_tree [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.662330] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338852, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.681710] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5240e65a-1964-44ba-edeb-2c12681a786f, 'name': SearchDatastore_Task, 'duration_secs': 0.0117} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.682628] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0adcb925-2607-4901-9ec4-20a37a11e1bb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.687780] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Waiting for the task: (returnval){ [ 839.687780] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524ab006-11db-4a41-5743-96451d1fc379" [ 839.687780] env[61243]: _type = "Task" [ 839.687780] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.695497] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524ab006-11db-4a41-5743-96451d1fc379, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.762774] env[61243]: DEBUG nova.compute.manager [req-402bb90d-0161-475c-8939-e3ca4e83f2c6 req-0398b1ef-2a24-4f65-bed9-54749f0a0f36 service nova] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Received event network-vif-plugged-f80550c9-ffa4-4f9f-a0c1-7be2562defa3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 839.763071] env[61243]: DEBUG oslo_concurrency.lockutils [req-402bb90d-0161-475c-8939-e3ca4e83f2c6 req-0398b1ef-2a24-4f65-bed9-54749f0a0f36 service nova] Acquiring lock "9a66fd05-1880-423a-8f60-4f11f1c75ab3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.763317] env[61243]: DEBUG oslo_concurrency.lockutils [req-402bb90d-0161-475c-8939-e3ca4e83f2c6 req-0398b1ef-2a24-4f65-bed9-54749f0a0f36 service nova] Lock "9a66fd05-1880-423a-8f60-4f11f1c75ab3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.763979] env[61243]: DEBUG oslo_concurrency.lockutils [req-402bb90d-0161-475c-8939-e3ca4e83f2c6 req-0398b1ef-2a24-4f65-bed9-54749f0a0f36 service nova] Lock "9a66fd05-1880-423a-8f60-4f11f1c75ab3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.763979] env[61243]: DEBUG nova.compute.manager [req-402bb90d-0161-475c-8939-e3ca4e83f2c6 req-0398b1ef-2a24-4f65-bed9-54749f0a0f36 service nova] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] No waiting events found dispatching network-vif-plugged-f80550c9-ffa4-4f9f-a0c1-7be2562defa3 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 839.764098] env[61243]: WARNING nova.compute.manager [req-402bb90d-0161-475c-8939-e3ca4e83f2c6 req-0398b1ef-2a24-4f65-bed9-54749f0a0f36 service nova] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Received unexpected event network-vif-plugged-f80550c9-ffa4-4f9f-a0c1-7be2562defa3 for instance with vm_state building and task_state spawning. [ 839.764284] env[61243]: DEBUG nova.compute.manager [req-402bb90d-0161-475c-8939-e3ca4e83f2c6 req-0398b1ef-2a24-4f65-bed9-54749f0a0f36 service nova] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Received event network-changed-f80550c9-ffa4-4f9f-a0c1-7be2562defa3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 839.764485] env[61243]: DEBUG nova.compute.manager [req-402bb90d-0161-475c-8939-e3ca4e83f2c6 req-0398b1ef-2a24-4f65-bed9-54749f0a0f36 service nova] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Refreshing instance network info cache due to event network-changed-f80550c9-ffa4-4f9f-a0c1-7be2562defa3. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 839.764711] env[61243]: DEBUG oslo_concurrency.lockutils [req-402bb90d-0161-475c-8939-e3ca4e83f2c6 req-0398b1ef-2a24-4f65-bed9-54749f0a0f36 service nova] Acquiring lock "refresh_cache-9a66fd05-1880-423a-8f60-4f11f1c75ab3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.897748] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338853, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.945681] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1338848, 'name': ReconfigVM_Task, 'duration_secs': 1.929705} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.946056] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba/99ce6ca3-a478-4ebe-bf1b-42459aacd9ba.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.946789] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d8b0b3b1-21ba-419f-83f2-9ecf7682172d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.953403] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 839.953403] env[61243]: value = "task-1338854" [ 839.953403] env[61243]: _type = "Task" [ 839.953403] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.961612] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1338854, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.003026] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Releasing lock "refresh_cache-9a66fd05-1880-423a-8f60-4f11f1c75ab3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.003026] env[61243]: DEBUG nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Instance network_info: |[{"id": "f80550c9-ffa4-4f9f-a0c1-7be2562defa3", "address": "fa:16:3e:83:38:f1", "network": {"id": "e10de432-a279-478d-a226-8cb1663f8f1e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-237966643-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39a655fc062e4322a5a129b66a45684f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf80550c9-ff", "ovs_interfaceid": "f80550c9-ffa4-4f9f-a0c1-7be2562defa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 840.003288] env[61243]: DEBUG oslo_concurrency.lockutils [req-402bb90d-0161-475c-8939-e3ca4e83f2c6 req-0398b1ef-2a24-4f65-bed9-54749f0a0f36 service nova] Acquired lock "refresh_cache-9a66fd05-1880-423a-8f60-4f11f1c75ab3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.003288] env[61243]: DEBUG nova.network.neutron [req-402bb90d-0161-475c-8939-e3ca4e83f2c6 req-0398b1ef-2a24-4f65-bed9-54749f0a0f36 service nova] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Refreshing network info cache for port f80550c9-ffa4-4f9f-a0c1-7be2562defa3 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 840.003288] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:38:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73f6629b-7f80-4a5b-8f15-c7a1635b3c33', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f80550c9-ffa4-4f9f-a0c1-7be2562defa3', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 840.020221] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Creating folder: Project (39a655fc062e4322a5a129b66a45684f). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 840.020994] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1c422f0e-64a6-4d36-b683-512c9e1d9912 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.035638] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Created folder: Project (39a655fc062e4322a5a129b66a45684f) in parent group-v285636. [ 840.036662] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Creating folder: Instances. Parent ref: group-v285687. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 840.036662] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2c547539-4045-442b-824c-0fd503867496 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.045643] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Created folder: Instances in parent group-v285687. [ 840.045963] env[61243]: DEBUG oslo.service.loopingcall [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.046206] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 840.046435] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-47c9b65f-3e95-4503-af1e-16d8c5c3341e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.063382] env[61243]: DEBUG oslo_vmware.rw_handles [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5251b8c9-0d7d-15c6-46d9-0399673adcca/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 840.064610] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47fafb44-11ff-4a0f-9070-584a0145cfe6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.070709] env[61243]: DEBUG oslo_vmware.rw_handles [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5251b8c9-0d7d-15c6-46d9-0399673adcca/disk-0.vmdk is in state: ready. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 840.070912] env[61243]: ERROR oslo_vmware.rw_handles [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5251b8c9-0d7d-15c6-46d9-0399673adcca/disk-0.vmdk due to incomplete transfer. [ 840.072171] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-589fe1d8-d070-46cf-a952-5b7e8d186e3b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.073630] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 840.073630] env[61243]: value = "task-1338857" [ 840.073630] env[61243]: _type = "Task" [ 840.073630] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.079190] env[61243]: DEBUG oslo_vmware.rw_handles [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5251b8c9-0d7d-15c6-46d9-0399673adcca/disk-0.vmdk. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 840.079190] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Uploaded image 137baf91-2436-47ab-9567-ee1d5ae5b3b2 to the Glance image server {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 840.085221] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Destroying the VM {{(pid=61243) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 840.085221] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-cb4c0425-d942-45a5-959c-275bd80440b2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.085956] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338857, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.090320] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 840.090320] env[61243]: value = "task-1338858" [ 840.090320] env[61243]: _type = "Task" [ 840.090320] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.099175] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338858, 'name': Destroy_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.135351] env[61243]: DEBUG nova.scheduler.client.report [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.166355] env[61243]: DEBUG oslo_vmware.api [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338852, 'name': PowerOnVM_Task, 'duration_secs': 0.701831} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.166827] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 840.167104] env[61243]: INFO nova.compute.manager [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Took 7.35 seconds to spawn the instance on the hypervisor. [ 840.167394] env[61243]: DEBUG nova.compute.manager [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.168200] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59bd68bb-d688-4027-84cd-679a2beac221 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.197654] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524ab006-11db-4a41-5743-96451d1fc379, 'name': SearchDatastore_Task, 'duration_secs': 0.010914} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.197933] env[61243]: DEBUG oslo_concurrency.lockutils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.198292] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 6f33b303-1ecd-478e-8fa3-33a5f9a5af50/6f33b303-1ecd-478e-8fa3-33a5f9a5af50.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 840.198517] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fa6b8d0c-5038-40c6-b77c-e54cc853f8ff {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.205185] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Waiting for the task: (returnval){ [ 840.205185] env[61243]: value = "task-1338859" [ 840.205185] env[61243]: _type = "Task" [ 840.205185] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.213574] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338859, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.245183] env[61243]: DEBUG nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 840.268624] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 840.268624] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 840.268624] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 840.268762] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 840.268762] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 840.268762] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 840.268999] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 840.269381] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 840.269381] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 840.271494] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 840.271494] env[61243]: DEBUG nova.virt.hardware [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 840.271494] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b8ec4e-ffae-4123-8d03-82f5f2acb2a4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.280793] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4aa357a-8a1d-4f61-8d38-bbd82bd92863 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.391918] env[61243]: DEBUG oslo_vmware.api [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1338853, 'name': PowerOnVM_Task, 'duration_secs': 0.527662} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.392284] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 840.395491] env[61243]: DEBUG nova.compute.manager [None req-4d17367a-43cf-4fcf-9e55-f7210c2145b3 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.396423] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ed3bc0-543a-422e-9419-6dc29f0df6a9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.465343] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1338854, 'name': Rename_Task, 'duration_secs': 0.136744} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.465343] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 840.465599] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea71d2a7-0fbb-4644-8a26-56c1c636337a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.472709] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 840.472709] env[61243]: value = "task-1338860" [ 840.472709] env[61243]: _type = "Task" [ 840.472709] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.481684] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1338860, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.589643] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338857, 'name': CreateVM_Task, 'duration_secs': 0.351651} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.589911] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 840.591189] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.591189] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.591671] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 840.592655] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59df9ae3-cc1b-4eb3-8a7b-ca653f235611 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.607126] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 840.607126] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f9395d-e43d-3a1a-f7af-07310755c441" [ 840.607126] env[61243]: _type = "Task" [ 840.607126] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.611312] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338858, 'name': Destroy_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.623688] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f9395d-e43d-3a1a-f7af-07310755c441, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.642610] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.642610] env[61243]: DEBUG nova.compute.manager [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 840.645472] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.136s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.648469] env[61243]: INFO nova.compute.claims [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 840.667629] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquiring lock "6cee3dd1-c13d-427d-889c-6cd3ae388a31" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.668043] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Lock "6cee3dd1-c13d-427d-889c-6cd3ae388a31" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.668346] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquiring lock "6cee3dd1-c13d-427d-889c-6cd3ae388a31-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.668958] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Lock "6cee3dd1-c13d-427d-889c-6cd3ae388a31-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.668958] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Lock "6cee3dd1-c13d-427d-889c-6cd3ae388a31-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.673132] env[61243]: INFO nova.compute.manager [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Terminating instance [ 840.677370] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquiring lock "refresh_cache-6cee3dd1-c13d-427d-889c-6cd3ae388a31" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.677370] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquired lock "refresh_cache-6cee3dd1-c13d-427d-889c-6cd3ae388a31" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.677370] env[61243]: DEBUG nova.network.neutron [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 840.685711] env[61243]: INFO nova.compute.manager [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Took 20.48 seconds to build instance. [ 840.722873] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338859, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.834775] env[61243]: DEBUG nova.network.neutron [req-402bb90d-0161-475c-8939-e3ca4e83f2c6 req-0398b1ef-2a24-4f65-bed9-54749f0a0f36 service nova] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Updated VIF entry in instance network info cache for port f80550c9-ffa4-4f9f-a0c1-7be2562defa3. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 840.835251] env[61243]: DEBUG nova.network.neutron [req-402bb90d-0161-475c-8939-e3ca4e83f2c6 req-0398b1ef-2a24-4f65-bed9-54749f0a0f36 service nova] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Updating instance_info_cache with network_info: [{"id": "f80550c9-ffa4-4f9f-a0c1-7be2562defa3", "address": "fa:16:3e:83:38:f1", "network": {"id": "e10de432-a279-478d-a226-8cb1663f8f1e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-237966643-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39a655fc062e4322a5a129b66a45684f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf80550c9-ff", "ovs_interfaceid": "f80550c9-ffa4-4f9f-a0c1-7be2562defa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.984642] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1338860, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.018049] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "36e55334-8628-4dd7-a845-f4ae3d8e7ff9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.018049] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "36e55334-8628-4dd7-a845-f4ae3d8e7ff9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.018049] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "36e55334-8628-4dd7-a845-f4ae3d8e7ff9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.018222] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "36e55334-8628-4dd7-a845-f4ae3d8e7ff9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.019194] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "36e55334-8628-4dd7-a845-f4ae3d8e7ff9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.021679] env[61243]: INFO nova.compute.manager [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Terminating instance [ 841.025768] env[61243]: DEBUG nova.network.neutron [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Successfully updated port: ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 841.032752] env[61243]: DEBUG nova.compute.manager [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 841.033298] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 841.034648] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a9a817f-4b43-48b8-ab72-dab62c56b557 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.047695] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 841.049711] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-06f2ecce-cc0e-4516-83eb-4b587ac2e753 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.103658] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338858, 'name': Destroy_Task, 'duration_secs': 0.717061} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.104073] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Destroyed the VM [ 841.104491] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Deleting Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 841.105578] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7192aeea-ae8f-42ad-8cdf-fd61fd093b11 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.112175] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 841.112175] env[61243]: value = "task-1338862" [ 841.112175] env[61243]: _type = "Task" [ 841.112175] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.127092] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f9395d-e43d-3a1a-f7af-07310755c441, 'name': SearchDatastore_Task, 'duration_secs': 0.057791} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.131353] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.131713] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 841.132050] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.132322] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.132708] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 841.132918] env[61243]: DEBUG oslo_vmware.api [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338862, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.133235] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 841.133497] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 841.133811] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Deleting the datastore file [datastore2] 36e55334-8628-4dd7-a845-f4ae3d8e7ff9 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.134132] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3960e01d-0735-4737-96fb-8c3e03cd9b2c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.136718] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3e99693e-6e47-43af-a298-a14dce570662 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.144204] env[61243]: DEBUG oslo_vmware.api [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 841.144204] env[61243]: value = "task-1338863" [ 841.144204] env[61243]: _type = "Task" [ 841.144204] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.146276] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 841.146464] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 841.150602] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7831b65-c8e2-47d7-98c1-48fd8077b799 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.154102] env[61243]: DEBUG nova.compute.utils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 841.158770] env[61243]: DEBUG nova.compute.manager [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 841.158770] env[61243]: DEBUG nova.network.neutron [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 841.164163] env[61243]: DEBUG oslo_vmware.api [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1338863, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.165462] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 841.165462] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cfbd29-8d15-4cd2-4350-d72bebd0f038" [ 841.165462] env[61243]: _type = "Task" [ 841.165462] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.173303] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cfbd29-8d15-4cd2-4350-d72bebd0f038, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.191858] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6bbd9c86-6a4f-4cbb-abb4-8e2ea88f1b34 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "96315d9e-4eda-4e3a-af0d-bdc52ab181e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.698s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.212282] env[61243]: DEBUG nova.network.neutron [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 841.220668] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338859, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.558499} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.220916] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 6f33b303-1ecd-478e-8fa3-33a5f9a5af50/6f33b303-1ecd-478e-8fa3-33a5f9a5af50.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 841.221187] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 841.221630] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2de2e1e6-befa-43f6-bbeb-acd8bc2e9897 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.228673] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Waiting for the task: (returnval){ [ 841.228673] env[61243]: value = "task-1338864" [ 841.228673] env[61243]: _type = "Task" [ 841.228673] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.240505] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338864, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.248226] env[61243]: DEBUG nova.policy [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a9e6dc85a31d4ec38efe0c750c49ead1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '113713170fb94b8a8d7985b9a02faf77', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 841.303125] env[61243]: DEBUG nova.network.neutron [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.338676] env[61243]: DEBUG oslo_concurrency.lockutils [req-402bb90d-0161-475c-8939-e3ca4e83f2c6 req-0398b1ef-2a24-4f65-bed9-54749f0a0f36 service nova] Releasing lock "refresh_cache-9a66fd05-1880-423a-8f60-4f11f1c75ab3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.428360] env[61243]: INFO nova.compute.manager [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Rebuilding instance [ 841.474337] env[61243]: DEBUG nova.compute.manager [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.474337] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b5a5af2-2272-4a4d-9bd9-d3018f41226e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.491745] env[61243]: DEBUG oslo_vmware.api [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1338860, 'name': PowerOnVM_Task, 'duration_secs': 0.602104} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.493564] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.493801] env[61243]: INFO nova.compute.manager [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Took 13.71 seconds to spawn the instance on the hypervisor. [ 841.494693] env[61243]: DEBUG nova.compute.manager [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.495031] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af18ab4d-a5ef-4982-aff6-b755b35dde3d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.542741] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "refresh_cache-1dc8ad6a-3a00-47c6-8985-481e7f1363cd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.543010] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquired lock "refresh_cache-1dc8ad6a-3a00-47c6-8985-481e7f1363cd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.543193] env[61243]: DEBUG nova.network.neutron [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 841.572574] env[61243]: DEBUG nova.network.neutron [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Successfully created port: 5ea2a618-21af-437b-b35b-69f2a38cc3dc {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.626845] env[61243]: DEBUG nova.compute.utils [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Cleaning up image 137baf91-2436-47ab-9567-ee1d5ae5b3b2 {{(pid=61243) delete_image /opt/stack/nova/nova/compute/utils.py:1322}} [ 841.658021] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] In vmwareapi:vmops:_destroy_instance, exception while deleting the VM contents from the disk: oslo_vmware.exceptions.CannotDeleteFileException: Cannot delete file [datastore2] 36e55334-8628-4dd7-a845-f4ae3d8e7ff9 [ 841.658021] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Traceback (most recent call last): [ 841.658021] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1113, in _destroy_instance [ 841.658021] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] ds_util.file_delete(self._session, [ 841.658021] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] File "/opt/stack/nova/nova/virt/vmwareapi/ds_util.py", line 219, in file_delete [ 841.658021] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] session._wait_for_task(file_delete_task) [ 841.658021] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task [ 841.658021] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] return self.wait_for_task(task_ref) [ 841.658021] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 398, in wait_for_task [ 841.658021] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] return evt.wait() [ 841.658021] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 841.658454] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] result = hub.switch() [ 841.658454] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 841.658454] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] return self.greenlet.switch() [ 841.658454] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner [ 841.658454] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] self.f(*self.args, **self.kw) [ 841.658454] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 448, in _poll_task [ 841.658454] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] raise exceptions.translate_fault(task_info.error) [ 841.658454] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] oslo_vmware.exceptions.CannotDeleteFileException: Cannot delete file [datastore2] 36e55334-8628-4dd7-a845-f4ae3d8e7ff9 [ 841.658454] env[61243]: ERROR nova.virt.vmwareapi.vmops [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] [ 841.658454] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 841.658777] env[61243]: INFO nova.compute.manager [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Took 0.62 seconds to destroy the instance on the hypervisor. [ 841.658777] env[61243]: DEBUG oslo.service.loopingcall [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.658777] env[61243]: DEBUG nova.compute.manager [-] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 841.658777] env[61243]: DEBUG nova.network.neutron [-] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 841.660522] env[61243]: DEBUG nova.compute.manager [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 841.675127] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "783d861c-8930-406e-8986-ef995de4c9dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.675127] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "783d861c-8930-406e-8986-ef995de4c9dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.682777] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cfbd29-8d15-4cd2-4350-d72bebd0f038, 'name': SearchDatastore_Task, 'duration_secs': 0.010874} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.684540] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-673deaaf-e8a5-4493-a242-7fd428be4ab8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.693421] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 841.693421] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ae9237-ff20-7764-7471-9ce7ba28831c" [ 841.693421] env[61243]: _type = "Task" [ 841.693421] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.694895] env[61243]: DEBUG nova.compute.manager [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 841.707123] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ae9237-ff20-7764-7471-9ce7ba28831c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.739575] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338864, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073164} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.739855] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.740834] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fc960f-70ef-4be8-a808-abdbdbd9c0b9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.771149] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 6f33b303-1ecd-478e-8fa3-33a5f9a5af50/6f33b303-1ecd-478e-8fa3-33a5f9a5af50.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.775593] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5a66f2b-ee4b-4dbc-ad47-94d2793c2864 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.796211] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Waiting for the task: (returnval){ [ 841.796211] env[61243]: value = "task-1338865" [ 841.796211] env[61243]: _type = "Task" [ 841.796211] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.807828] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Releasing lock "refresh_cache-6cee3dd1-c13d-427d-889c-6cd3ae388a31" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.808334] env[61243]: DEBUG nova.compute.manager [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 841.808564] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 841.809864] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338865, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.809864] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3cbe45-44d3-4eca-aa40-bca56fa4243e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.817643] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 841.817927] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc926cbc-b968-4613-8a85-790332eeff52 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.824020] env[61243]: DEBUG oslo_vmware.api [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 841.824020] env[61243]: value = "task-1338866" [ 841.824020] env[61243]: _type = "Task" [ 841.824020] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.833018] env[61243]: DEBUG oslo_vmware.api [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338866, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.845353] env[61243]: DEBUG nova.compute.manager [req-c0aa4101-afed-44e4-8594-9273cc0f6594 req-2339a55e-12bd-436c-bdc8-2fbc2d33d32d service nova] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Received event network-vif-plugged-ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.845619] env[61243]: DEBUG oslo_concurrency.lockutils [req-c0aa4101-afed-44e4-8594-9273cc0f6594 req-2339a55e-12bd-436c-bdc8-2fbc2d33d32d service nova] Acquiring lock "1dc8ad6a-3a00-47c6-8985-481e7f1363cd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.845839] env[61243]: DEBUG oslo_concurrency.lockutils [req-c0aa4101-afed-44e4-8594-9273cc0f6594 req-2339a55e-12bd-436c-bdc8-2fbc2d33d32d service nova] Lock "1dc8ad6a-3a00-47c6-8985-481e7f1363cd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.846069] env[61243]: DEBUG oslo_concurrency.lockutils [req-c0aa4101-afed-44e4-8594-9273cc0f6594 req-2339a55e-12bd-436c-bdc8-2fbc2d33d32d service nova] Lock "1dc8ad6a-3a00-47c6-8985-481e7f1363cd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.846615] env[61243]: DEBUG nova.compute.manager [req-c0aa4101-afed-44e4-8594-9273cc0f6594 req-2339a55e-12bd-436c-bdc8-2fbc2d33d32d service nova] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] No waiting events found dispatching network-vif-plugged-ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 841.846615] env[61243]: WARNING nova.compute.manager [req-c0aa4101-afed-44e4-8594-9273cc0f6594 req-2339a55e-12bd-436c-bdc8-2fbc2d33d32d service nova] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Received unexpected event network-vif-plugged-ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6 for instance with vm_state building and task_state spawning. [ 841.846615] env[61243]: DEBUG nova.compute.manager [req-c0aa4101-afed-44e4-8594-9273cc0f6594 req-2339a55e-12bd-436c-bdc8-2fbc2d33d32d service nova] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Received event network-changed-ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.846792] env[61243]: DEBUG nova.compute.manager [req-c0aa4101-afed-44e4-8594-9273cc0f6594 req-2339a55e-12bd-436c-bdc8-2fbc2d33d32d service nova] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Refreshing instance network info cache due to event network-changed-ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 841.847065] env[61243]: DEBUG oslo_concurrency.lockutils [req-c0aa4101-afed-44e4-8594-9273cc0f6594 req-2339a55e-12bd-436c-bdc8-2fbc2d33d32d service nova] Acquiring lock "refresh_cache-1dc8ad6a-3a00-47c6-8985-481e7f1363cd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.994500] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 841.994785] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-77d6c237-43cb-4a0e-8a1a-70dbd602fa4c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.001906] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 842.001906] env[61243]: value = "task-1338867" [ 842.001906] env[61243]: _type = "Task" [ 842.001906] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.021123] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338867, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.024381] env[61243]: INFO nova.compute.manager [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Took 34.79 seconds to build instance. [ 842.055744] env[61243]: DEBUG oslo_concurrency.lockutils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "6036199d-cda9-42f8-8669-bc099d2b5ffe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.056209] env[61243]: DEBUG oslo_concurrency.lockutils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "6036199d-cda9-42f8-8669-bc099d2b5ffe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.087992] env[61243]: DEBUG nova.network.neutron [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 842.095540] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8a1e692-023f-43a5-b3ca-048b91091422 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.103035] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3829b7c7-2b3f-45be-b522-db8fee044703 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.140695] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eadf5e3a-00c7-4ab3-98e4-c0d17c43c1f7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.150684] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1cf05cf-46e7-4637-8f73-d8e342b158ec {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.166821] env[61243]: DEBUG nova.compute.provider_tree [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.206068] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ae9237-ff20-7764-7471-9ce7ba28831c, 'name': SearchDatastore_Task, 'duration_secs': 0.058693} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.211059] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.211059] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 9a66fd05-1880-423a-8f60-4f11f1c75ab3/9a66fd05-1880-423a-8f60-4f11f1c75ab3.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 842.211846] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9b8de7ff-9a31-4c8f-87cc-2b5f9cb71e75 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.219114] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 842.219114] env[61243]: value = "task-1338868" [ 842.219114] env[61243]: _type = "Task" [ 842.219114] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.223771] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.227150] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338868, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.296937] env[61243]: DEBUG nova.network.neutron [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Updating instance_info_cache with network_info: [{"id": "ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6", "address": "fa:16:3e:db:ed:3d", "network": {"id": "e10de432-a279-478d-a226-8cb1663f8f1e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-237966643-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39a655fc062e4322a5a129b66a45684f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebf3bb9d-c4", "ovs_interfaceid": "ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.307179] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338865, 'name': ReconfigVM_Task, 'duration_secs': 0.452879} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.307513] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 6f33b303-1ecd-478e-8fa3-33a5f9a5af50/6f33b303-1ecd-478e-8fa3-33a5f9a5af50.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.308156] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8e8fc7c3-b987-41ca-9147-98ff8e74ef77 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.314424] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Waiting for the task: (returnval){ [ 842.314424] env[61243]: value = "task-1338869" [ 842.314424] env[61243]: _type = "Task" [ 842.314424] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.323201] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338869, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.331809] env[61243]: DEBUG oslo_vmware.api [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338866, 'name': PowerOffVM_Task, 'duration_secs': 0.223713} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.333038] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 842.333038] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 842.333038] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c0cba353-98ae-4839-808d-4e057341efef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.354873] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 842.355120] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 842.355352] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Deleting the datastore file [datastore1] 6cee3dd1-c13d-427d-889c-6cd3ae388a31 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 842.355627] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d19cfd87-d03d-4d2e-96c4-1dc257f5ac8b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.361650] env[61243]: DEBUG oslo_vmware.api [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for the task: (returnval){ [ 842.361650] env[61243]: value = "task-1338871" [ 842.361650] env[61243]: _type = "Task" [ 842.361650] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.369142] env[61243]: DEBUG oslo_vmware.api [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338871, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.515270] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338867, 'name': PowerOffVM_Task, 'duration_secs': 0.17584} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.515500] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 842.515728] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 842.516531] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81cbd46c-6b0b-47a7-9079-80cb5261ccd2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.523247] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 842.523499] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6d43d34e-b95b-46a3-b9f2-b15bcfdc0a93 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.527619] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cac53eb9-b68b-4e6a-b646-9a028b2ce909 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.108s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.546638] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 842.546747] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 842.547317] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Deleting the datastore file [datastore2] 96315d9e-4eda-4e3a-af0d-bdc52ab181e3 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 842.547317] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bdc1055d-ee34-4ddd-b8a3-7bfdafe55cf0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.555290] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 842.555290] env[61243]: value = "task-1338873" [ 842.555290] env[61243]: _type = "Task" [ 842.555290] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.565533] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338873, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.644642] env[61243]: DEBUG nova.network.neutron [-] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.670689] env[61243]: DEBUG nova.scheduler.client.report [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.677058] env[61243]: DEBUG nova.compute.manager [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 842.704050] env[61243]: DEBUG nova.virt.hardware [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 842.704050] env[61243]: DEBUG nova.virt.hardware [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 842.704367] env[61243]: DEBUG nova.virt.hardware [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.704714] env[61243]: DEBUG nova.virt.hardware [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 842.704994] env[61243]: DEBUG nova.virt.hardware [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.705373] env[61243]: DEBUG nova.virt.hardware [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 842.706763] env[61243]: DEBUG nova.virt.hardware [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 842.706763] env[61243]: DEBUG nova.virt.hardware [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 842.706763] env[61243]: DEBUG nova.virt.hardware [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 842.706763] env[61243]: DEBUG nova.virt.hardware [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 842.706763] env[61243]: DEBUG nova.virt.hardware [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 842.708102] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d6f6ce-b004-4d32-89e4-938e905bb26e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.718865] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0930384c-9aa1-44f9-8909-ba2292df8168 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.732984] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338868, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.803032] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Releasing lock "refresh_cache-1dc8ad6a-3a00-47c6-8985-481e7f1363cd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.803032] env[61243]: DEBUG nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Instance network_info: |[{"id": "ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6", "address": "fa:16:3e:db:ed:3d", "network": {"id": "e10de432-a279-478d-a226-8cb1663f8f1e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-237966643-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39a655fc062e4322a5a129b66a45684f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebf3bb9d-c4", "ovs_interfaceid": "ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 842.803794] env[61243]: DEBUG oslo_concurrency.lockutils [req-c0aa4101-afed-44e4-8594-9273cc0f6594 req-2339a55e-12bd-436c-bdc8-2fbc2d33d32d service nova] Acquired lock "refresh_cache-1dc8ad6a-3a00-47c6-8985-481e7f1363cd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.803794] env[61243]: DEBUG nova.network.neutron [req-c0aa4101-afed-44e4-8594-9273cc0f6594 req-2339a55e-12bd-436c-bdc8-2fbc2d33d32d service nova] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Refreshing network info cache for port ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 842.804579] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:ed:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73f6629b-7f80-4a5b-8f15-c7a1635b3c33', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.812428] env[61243]: DEBUG oslo.service.loopingcall [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.813599] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 842.813832] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a90cd149-7c00-4420-bc00-f26891032bf6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.842437] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338869, 'name': Rename_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.844042] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.844042] env[61243]: value = "task-1338874" [ 842.844042] env[61243]: _type = "Task" [ 842.844042] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.853908] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338874, 'name': CreateVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.872823] env[61243]: DEBUG oslo_vmware.api [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Task: {'id': task-1338871, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.266027} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.873153] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 842.873355] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 842.873542] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 842.873854] env[61243]: INFO nova.compute.manager [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Took 1.07 seconds to destroy the instance on the hypervisor. [ 842.874112] env[61243]: DEBUG oslo.service.loopingcall [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.874352] env[61243]: DEBUG nova.compute.manager [-] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 842.874408] env[61243]: DEBUG nova.network.neutron [-] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 842.906787] env[61243]: DEBUG nova.network.neutron [-] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 843.030931] env[61243]: DEBUG nova.compute.manager [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.064727] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338873, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.42921} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.067510] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 843.067510] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 843.067510] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 843.147714] env[61243]: INFO nova.compute.manager [-] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Took 1.49 seconds to deallocate network for instance. [ 843.153695] env[61243]: DEBUG oslo_concurrency.lockutils [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.175448] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.530s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.176956] env[61243]: DEBUG nova.compute.manager [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 843.179451] env[61243]: DEBUG oslo_concurrency.lockutils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.095s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.180246] env[61243]: INFO nova.compute.claims [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.239762] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338868, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.757832} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.240094] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 9a66fd05-1880-423a-8f60-4f11f1c75ab3/9a66fd05-1880-423a-8f60-4f11f1c75ab3.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 843.240307] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 843.240539] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0fa64ae8-f78c-4a36-bfdd-5e97f7279c99 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.248213] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 843.248213] env[61243]: value = "task-1338875" [ 843.248213] env[61243]: _type = "Task" [ 843.248213] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.258883] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338875, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.341016] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338869, 'name': Rename_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.353974] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338874, 'name': CreateVM_Task, 'duration_secs': 0.377087} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.354212] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 843.355126] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.355362] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.355802] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.356335] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d97299e-9997-475f-babe-08275a923d55 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.361257] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 843.361257] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]526ba314-48e1-6a21-099d-7e4976d34d1d" [ 843.361257] env[61243]: _type = "Task" [ 843.361257] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.369398] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]526ba314-48e1-6a21-099d-7e4976d34d1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.408724] env[61243]: DEBUG nova.network.neutron [-] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.508859] env[61243]: DEBUG nova.network.neutron [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Successfully updated port: 5ea2a618-21af-437b-b35b-69f2a38cc3dc {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 843.563685] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.617936] env[61243]: DEBUG nova.network.neutron [req-c0aa4101-afed-44e4-8594-9273cc0f6594 req-2339a55e-12bd-436c-bdc8-2fbc2d33d32d service nova] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Updated VIF entry in instance network info cache for port ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 843.618309] env[61243]: DEBUG nova.network.neutron [req-c0aa4101-afed-44e4-8594-9273cc0f6594 req-2339a55e-12bd-436c-bdc8-2fbc2d33d32d service nova] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Updating instance_info_cache with network_info: [{"id": "ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6", "address": "fa:16:3e:db:ed:3d", "network": {"id": "e10de432-a279-478d-a226-8cb1663f8f1e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-237966643-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39a655fc062e4322a5a129b66a45684f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebf3bb9d-c4", "ovs_interfaceid": "ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.655716] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.686371] env[61243]: DEBUG nova.compute.utils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.689481] env[61243]: DEBUG nova.compute.manager [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 843.690089] env[61243]: DEBUG nova.network.neutron [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 843.729560] env[61243]: DEBUG nova.policy [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '51057f2af9244ee5a0a00cd159f622a0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '93625b58c8674dc0980e3d9fee62ab18', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 843.757577] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338875, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091023} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.758223] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 843.760251] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-530308e3-f2ed-485f-ad39-8732d56e1d63 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.782056] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 9a66fd05-1880-423a-8f60-4f11f1c75ab3/9a66fd05-1880-423a-8f60-4f11f1c75ab3.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 843.782428] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad3199a7-53d6-4596-894c-9c4303041f24 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.801672] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 843.801672] env[61243]: value = "task-1338876" [ 843.801672] env[61243]: _type = "Task" [ 843.801672] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.809754] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338876, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.840159] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338869, 'name': Rename_Task, 'duration_secs': 1.197683} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.841034] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 843.841154] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-036fde51-29b6-44fe-be8b-7cfda68da8a2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.846879] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Waiting for the task: (returnval){ [ 843.846879] env[61243]: value = "task-1338877" [ 843.846879] env[61243]: _type = "Task" [ 843.846879] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.854711] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338877, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.874028] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]526ba314-48e1-6a21-099d-7e4976d34d1d, 'name': SearchDatastore_Task, 'duration_secs': 0.037203} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.876017] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.876017] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.876017] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.876017] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.876214] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.876214] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-692a4ecd-70ed-4ab8-bc00-cb9b43653ff5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.884562] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.884870] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 843.885791] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4351012d-b274-46b3-a421-8771392be40e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.891951] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 843.891951] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c4732e-4c4b-9703-aaea-e992c487f95c" [ 843.891951] env[61243]: _type = "Task" [ 843.891951] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.899619] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c4732e-4c4b-9703-aaea-e992c487f95c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.911311] env[61243]: INFO nova.compute.manager [-] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Took 1.04 seconds to deallocate network for instance. [ 844.012306] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "refresh_cache-ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.012701] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "refresh_cache-ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.012701] env[61243]: DEBUG nova.network.neutron [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 844.058165] env[61243]: DEBUG nova.network.neutron [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Successfully created port: 817c2755-c3a0-4de3-9427-322d92ba1d30 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 844.090947] env[61243]: DEBUG nova.compute.manager [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Received event network-vif-deleted-471f4f83-e6c9-47ce-bfde-c78219c698a6 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.092026] env[61243]: DEBUG nova.compute.manager [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Received event network-vif-plugged-5ea2a618-21af-437b-b35b-69f2a38cc3dc {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.092026] env[61243]: DEBUG oslo_concurrency.lockutils [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] Acquiring lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.092026] env[61243]: DEBUG oslo_concurrency.lockutils [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] Lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.092236] env[61243]: DEBUG oslo_concurrency.lockutils [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] Lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.092276] env[61243]: DEBUG nova.compute.manager [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] No waiting events found dispatching network-vif-plugged-5ea2a618-21af-437b-b35b-69f2a38cc3dc {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 844.092441] env[61243]: WARNING nova.compute.manager [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Received unexpected event network-vif-plugged-5ea2a618-21af-437b-b35b-69f2a38cc3dc for instance with vm_state building and task_state spawning. [ 844.092607] env[61243]: DEBUG nova.compute.manager [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Received event network-changed-b22c18f9-e3a1-4c8b-a26a-17c651ca16b5 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.092765] env[61243]: DEBUG nova.compute.manager [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Refreshing instance network info cache due to event network-changed-b22c18f9-e3a1-4c8b-a26a-17c651ca16b5. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 844.092957] env[61243]: DEBUG oslo_concurrency.lockutils [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] Acquiring lock "refresh_cache-99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.093120] env[61243]: DEBUG oslo_concurrency.lockutils [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] Acquired lock "refresh_cache-99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.093266] env[61243]: DEBUG nova.network.neutron [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Refreshing network info cache for port b22c18f9-e3a1-4c8b-a26a-17c651ca16b5 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.104722] env[61243]: DEBUG nova.virt.hardware [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.104967] env[61243]: DEBUG nova.virt.hardware [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.105677] env[61243]: DEBUG nova.virt.hardware [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.105677] env[61243]: DEBUG nova.virt.hardware [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.105677] env[61243]: DEBUG nova.virt.hardware [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.105677] env[61243]: DEBUG nova.virt.hardware [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.105974] env[61243]: DEBUG nova.virt.hardware [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.106100] env[61243]: DEBUG nova.virt.hardware [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.106262] env[61243]: DEBUG nova.virt.hardware [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.106440] env[61243]: DEBUG nova.virt.hardware [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.106644] env[61243]: DEBUG nova.virt.hardware [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.107823] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93daec1-bd52-4a6e-bd9e-3003d8a5ac96 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.117307] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a969ac-0241-40e5-9748-8463540d3a23 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.121852] env[61243]: DEBUG oslo_concurrency.lockutils [req-c0aa4101-afed-44e4-8594-9273cc0f6594 req-2339a55e-12bd-436c-bdc8-2fbc2d33d32d service nova] Releasing lock "refresh_cache-1dc8ad6a-3a00-47c6-8985-481e7f1363cd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.140648] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Instance VIF info [] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.150027] env[61243]: DEBUG oslo.service.loopingcall [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.150027] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 844.150027] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6d269ea1-02ed-419f-80e6-46d1b8c045c6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.170415] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.170415] env[61243]: value = "task-1338878" [ 844.170415] env[61243]: _type = "Task" [ 844.170415] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.179581] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338878, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.192678] env[61243]: DEBUG nova.compute.manager [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 844.316915] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338876, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.361214] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338877, 'name': PowerOnVM_Task} progress is 71%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.403530] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c4732e-4c4b-9703-aaea-e992c487f95c, 'name': SearchDatastore_Task, 'duration_secs': 0.010897} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.404754] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95f65536-6bc7-4e16-a9a8-2bed6cf10cab {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.413388] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 844.413388] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cd8145-ca87-44b4-c9da-e66faf9d4b6d" [ 844.413388] env[61243]: _type = "Task" [ 844.413388] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.423885] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.424682] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cd8145-ca87-44b4-c9da-e66faf9d4b6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.523630] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e26c2ee-b6ed-4b2b-9db1-8899fbd956a5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.530932] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cdfe82c-8836-4579-846b-bbd38d8802d5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.564099] env[61243]: DEBUG nova.network.neutron [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.567562] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c715dd6-4ac2-4a08-affa-cb175a3bbf0f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.574863] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b878504-f253-4048-91ed-499f642106a4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.588492] env[61243]: DEBUG nova.compute.provider_tree [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.681446] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338878, 'name': CreateVM_Task, 'duration_secs': 0.441394} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.684922] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 844.686140] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.686140] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.686140] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.686465] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bd90f08-10dc-4206-9ca1-45edeb12d389 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.691508] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 844.691508] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523e5818-e61b-eb93-64f1-2edeade6f5de" [ 844.691508] env[61243]: _type = "Task" [ 844.691508] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.702603] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523e5818-e61b-eb93-64f1-2edeade6f5de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.809361] env[61243]: DEBUG nova.network.neutron [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Updating instance_info_cache with network_info: [{"id": "5ea2a618-21af-437b-b35b-69f2a38cc3dc", "address": "fa:16:3e:6f:f0:da", "network": {"id": "0788f9ac-5a15-4431-a711-9b4f6cf79fbc", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1210957657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "113713170fb94b8a8d7985b9a02faf77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c7821ea-f92f-4f06-a4cb-05e1186a9d22", "external-id": "nsx-vlan-transportzone-69", "segmentation_id": 69, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ea2a618-21", "ovs_interfaceid": "5ea2a618-21af-437b-b35b-69f2a38cc3dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.816516] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338876, 'name': ReconfigVM_Task, 'duration_secs': 0.792613} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.816797] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 9a66fd05-1880-423a-8f60-4f11f1c75ab3/9a66fd05-1880-423a-8f60-4f11f1c75ab3.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 844.817773] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd9f3581-33f2-42ef-b894-adb6ac022f95 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.835917] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 844.835917] env[61243]: value = "task-1338879" [ 844.835917] env[61243]: _type = "Task" [ 844.835917] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.844469] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338879, 'name': Rename_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.858801] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338877, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.902342] env[61243]: DEBUG nova.network.neutron [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Updated VIF entry in instance network info cache for port b22c18f9-e3a1-4c8b-a26a-17c651ca16b5. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 844.902810] env[61243]: DEBUG nova.network.neutron [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Updating instance_info_cache with network_info: [{"id": "b22c18f9-e3a1-4c8b-a26a-17c651ca16b5", "address": "fa:16:3e:55:c6:a8", "network": {"id": "523d4cf1-467c-4ea6-80a4-ebbd01016efa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-919325376-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7651b565d1647eb948fe07faba32a41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb22c18f9-e3", "ovs_interfaceid": "b22c18f9-e3a1-4c8b-a26a-17c651ca16b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.924512] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cd8145-ca87-44b4-c9da-e66faf9d4b6d, 'name': SearchDatastore_Task, 'duration_secs': 0.01585} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.924776] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.925063] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 1dc8ad6a-3a00-47c6-8985-481e7f1363cd/1dc8ad6a-3a00-47c6-8985-481e7f1363cd.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 844.925351] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-449f100e-9cab-4ff6-b9b5-ec0bcb2fe380 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.931386] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 844.931386] env[61243]: value = "task-1338880" [ 844.931386] env[61243]: _type = "Task" [ 844.931386] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.939901] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338880, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.094035] env[61243]: DEBUG nova.scheduler.client.report [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.201925] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523e5818-e61b-eb93-64f1-2edeade6f5de, 'name': SearchDatastore_Task, 'duration_secs': 0.047588} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.202295] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.202510] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.202754] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.202912] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.203116] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.204314] env[61243]: DEBUG nova.compute.manager [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 845.206240] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c25ac1bd-9e6c-4beb-ba76-1d3e564f6df6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.217746] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.217883] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 845.218783] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06de29f0-9d62-460c-a45f-9e06188183a8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.227682] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 845.227682] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e3bd71-ad7f-3a5a-3ed0-21cbd74cc99e" [ 845.227682] env[61243]: _type = "Task" [ 845.227682] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.232562] env[61243]: DEBUG nova.virt.hardware [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 845.232801] env[61243]: DEBUG nova.virt.hardware [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 845.232980] env[61243]: DEBUG nova.virt.hardware [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 845.233209] env[61243]: DEBUG nova.virt.hardware [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 845.233368] env[61243]: DEBUG nova.virt.hardware [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 845.233538] env[61243]: DEBUG nova.virt.hardware [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 845.233751] env[61243]: DEBUG nova.virt.hardware [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 845.233918] env[61243]: DEBUG nova.virt.hardware [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 845.234130] env[61243]: DEBUG nova.virt.hardware [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 845.234323] env[61243]: DEBUG nova.virt.hardware [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 845.234504] env[61243]: DEBUG nova.virt.hardware [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 845.235293] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55cbe87b-34c7-48b2-8ed5-b6edfd3a91b0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.243087] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e3bd71-ad7f-3a5a-3ed0-21cbd74cc99e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.246155] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1455b126-f55d-42ea-b612-3416b43afb9d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.312911] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "refresh_cache-ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.313334] env[61243]: DEBUG nova.compute.manager [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Instance network_info: |[{"id": "5ea2a618-21af-437b-b35b-69f2a38cc3dc", "address": "fa:16:3e:6f:f0:da", "network": {"id": "0788f9ac-5a15-4431-a711-9b4f6cf79fbc", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1210957657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "113713170fb94b8a8d7985b9a02faf77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c7821ea-f92f-4f06-a4cb-05e1186a9d22", "external-id": "nsx-vlan-transportzone-69", "segmentation_id": 69, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ea2a618-21", "ovs_interfaceid": "5ea2a618-21af-437b-b35b-69f2a38cc3dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 845.313819] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:f0:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5c7821ea-f92f-4f06-a4cb-05e1186a9d22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5ea2a618-21af-437b-b35b-69f2a38cc3dc', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 845.321362] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Creating folder: Project (113713170fb94b8a8d7985b9a02faf77). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 845.321678] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e4629b72-3648-4d15-914f-32091a58e540 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.335843] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Created folder: Project (113713170fb94b8a8d7985b9a02faf77) in parent group-v285636. [ 845.336021] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Creating folder: Instances. Parent ref: group-v285692. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 845.336358] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7b940d28-75aa-4f6c-be20-c7d63cab1d32 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.347801] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338879, 'name': Rename_Task, 'duration_secs': 0.221836} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.348075] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 845.349242] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d2389bca-6400-4b17-af37-963b2b5eeea9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.350757] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Created folder: Instances in parent group-v285692. [ 845.350982] env[61243]: DEBUG oslo.service.loopingcall [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.353677] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 845.354526] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b09dc38f-51ff-41d2-977c-e7c89b2622de {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.370683] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 845.370683] env[61243]: value = "task-1338883" [ 845.370683] env[61243]: _type = "Task" [ 845.370683] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.378301] env[61243]: DEBUG oslo_vmware.api [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338877, 'name': PowerOnVM_Task, 'duration_secs': 1.091016} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.379527] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 845.379768] env[61243]: INFO nova.compute.manager [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Took 10.09 seconds to spawn the instance on the hypervisor. [ 845.379984] env[61243]: DEBUG nova.compute.manager [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 845.380268] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 845.380268] env[61243]: value = "task-1338884" [ 845.380268] env[61243]: _type = "Task" [ 845.380268] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.381012] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde2115a-4f20-4865-8735-eb65d6b76e4c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.390813] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338883, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.396190] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338884, 'name': CreateVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.405938] env[61243]: DEBUG oslo_concurrency.lockutils [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] Releasing lock "refresh_cache-99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.406241] env[61243]: DEBUG nova.compute.manager [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Received event network-changed-5ea2a618-21af-437b-b35b-69f2a38cc3dc {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.406429] env[61243]: DEBUG nova.compute.manager [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Refreshing instance network info cache due to event network-changed-5ea2a618-21af-437b-b35b-69f2a38cc3dc. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 845.406648] env[61243]: DEBUG oslo_concurrency.lockutils [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] Acquiring lock "refresh_cache-ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.406797] env[61243]: DEBUG oslo_concurrency.lockutils [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] Acquired lock "refresh_cache-ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.406958] env[61243]: DEBUG nova.network.neutron [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Refreshing network info cache for port 5ea2a618-21af-437b-b35b-69f2a38cc3dc {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 845.444629] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338880, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.602023] env[61243]: DEBUG oslo_concurrency.lockutils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.602023] env[61243]: DEBUG nova.compute.manager [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 845.603707] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.173s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.605820] env[61243]: INFO nova.compute.claims [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.618802] env[61243]: DEBUG nova.network.neutron [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Successfully updated port: 817c2755-c3a0-4de3-9427-322d92ba1d30 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 845.738586] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e3bd71-ad7f-3a5a-3ed0-21cbd74cc99e, 'name': SearchDatastore_Task, 'duration_secs': 0.019583} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.739433] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc0457ca-d0a0-415a-abeb-04c23dc9f878 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.744963] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 845.744963] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5262675d-ef63-5d06-2ce0-6139e888e067" [ 845.744963] env[61243]: _type = "Task" [ 845.744963] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.752633] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5262675d-ef63-5d06-2ce0-6139e888e067, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.881158] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338883, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.892494] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338884, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.907405] env[61243]: INFO nova.compute.manager [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Took 24.95 seconds to build instance. [ 845.944077] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338880, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.806389} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.946747] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 1dc8ad6a-3a00-47c6-8985-481e7f1363cd/1dc8ad6a-3a00-47c6-8985-481e7f1363cd.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 845.946979] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.947318] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf13841a-b49b-4b48-ba8a-1c8d03f6adc7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.955783] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 845.955783] env[61243]: value = "task-1338885" [ 845.955783] env[61243]: _type = "Task" [ 845.955783] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.964285] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338885, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.110349] env[61243]: DEBUG nova.compute.utils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.112035] env[61243]: DEBUG nova.compute.manager [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 846.112713] env[61243]: DEBUG nova.network.neutron [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 846.128399] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Acquiring lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.129256] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Acquired lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.129256] env[61243]: DEBUG nova.network.neutron [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 846.132208] env[61243]: DEBUG nova.compute.manager [req-d89132e3-2c52-4474-abec-631981a52ccc req-6e0830d0-a5c1-451d-adf4-9cb0ed363372 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Received event network-vif-plugged-817c2755-c3a0-4de3-9427-322d92ba1d30 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.132897] env[61243]: DEBUG oslo_concurrency.lockutils [req-d89132e3-2c52-4474-abec-631981a52ccc req-6e0830d0-a5c1-451d-adf4-9cb0ed363372 service nova] Acquiring lock "36db1ed5-846f-4ad6-8cee-38b73ff00321-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.133141] env[61243]: DEBUG oslo_concurrency.lockutils [req-d89132e3-2c52-4474-abec-631981a52ccc req-6e0830d0-a5c1-451d-adf4-9cb0ed363372 service nova] Lock "36db1ed5-846f-4ad6-8cee-38b73ff00321-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.133323] env[61243]: DEBUG oslo_concurrency.lockutils [req-d89132e3-2c52-4474-abec-631981a52ccc req-6e0830d0-a5c1-451d-adf4-9cb0ed363372 service nova] Lock "36db1ed5-846f-4ad6-8cee-38b73ff00321-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.133495] env[61243]: DEBUG nova.compute.manager [req-d89132e3-2c52-4474-abec-631981a52ccc req-6e0830d0-a5c1-451d-adf4-9cb0ed363372 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] No waiting events found dispatching network-vif-plugged-817c2755-c3a0-4de3-9427-322d92ba1d30 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 846.136071] env[61243]: WARNING nova.compute.manager [req-d89132e3-2c52-4474-abec-631981a52ccc req-6e0830d0-a5c1-451d-adf4-9cb0ed363372 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Received unexpected event network-vif-plugged-817c2755-c3a0-4de3-9427-322d92ba1d30 for instance with vm_state building and task_state spawning. [ 846.136071] env[61243]: DEBUG nova.compute.manager [req-d89132e3-2c52-4474-abec-631981a52ccc req-6e0830d0-a5c1-451d-adf4-9cb0ed363372 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Received event network-changed-817c2755-c3a0-4de3-9427-322d92ba1d30 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.136071] env[61243]: DEBUG nova.compute.manager [req-d89132e3-2c52-4474-abec-631981a52ccc req-6e0830d0-a5c1-451d-adf4-9cb0ed363372 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Refreshing instance network info cache due to event network-changed-817c2755-c3a0-4de3-9427-322d92ba1d30. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 846.136071] env[61243]: DEBUG oslo_concurrency.lockutils [req-d89132e3-2c52-4474-abec-631981a52ccc req-6e0830d0-a5c1-451d-adf4-9cb0ed363372 service nova] Acquiring lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.179204] env[61243]: DEBUG nova.network.neutron [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Updated VIF entry in instance network info cache for port 5ea2a618-21af-437b-b35b-69f2a38cc3dc. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 846.179546] env[61243]: DEBUG nova.network.neutron [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Updating instance_info_cache with network_info: [{"id": "5ea2a618-21af-437b-b35b-69f2a38cc3dc", "address": "fa:16:3e:6f:f0:da", "network": {"id": "0788f9ac-5a15-4431-a711-9b4f6cf79fbc", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1210957657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "113713170fb94b8a8d7985b9a02faf77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c7821ea-f92f-4f06-a4cb-05e1186a9d22", "external-id": "nsx-vlan-transportzone-69", "segmentation_id": 69, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ea2a618-21", "ovs_interfaceid": "5ea2a618-21af-437b-b35b-69f2a38cc3dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.190037] env[61243]: DEBUG nova.network.neutron [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 846.202850] env[61243]: DEBUG nova.policy [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ecfff92a554145c0bd024e59fb3d6f52', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66e2ba61b3354ef1b496232264d27600', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 846.257112] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5262675d-ef63-5d06-2ce0-6139e888e067, 'name': SearchDatastore_Task, 'duration_secs': 0.015506} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.262330] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.262330] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 96315d9e-4eda-4e3a-af0d-bdc52ab181e3/96315d9e-4eda-4e3a-af0d-bdc52ab181e3.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 846.262330] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-27bb4978-3963-4c64-87e4-bd8b2fb83038 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.276649] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 846.276649] env[61243]: value = "task-1338886" [ 846.276649] env[61243]: _type = "Task" [ 846.276649] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.287353] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338886, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.381487] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338883, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.392650] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338884, 'name': CreateVM_Task, 'duration_secs': 0.876854} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.392813] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 846.393420] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.393602] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.393906] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 846.394190] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c3e79c4-94f6-4b86-bb0e-4eaa8083b73b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.398439] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 846.398439] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f50975-8b5f-7a45-157d-3c2621b521f5" [ 846.398439] env[61243]: _type = "Task" [ 846.398439] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.407351] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f50975-8b5f-7a45-157d-3c2621b521f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.408272] env[61243]: DEBUG nova.network.neutron [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Updating instance_info_cache with network_info: [{"id": "817c2755-c3a0-4de3-9427-322d92ba1d30", "address": "fa:16:3e:5c:f9:55", "network": {"id": "f6dda51c-55ed-4b80-b54d-ade24626e5fd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1249632116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93625b58c8674dc0980e3d9fee62ab18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap817c2755-c3", "ovs_interfaceid": "817c2755-c3a0-4de3-9427-322d92ba1d30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.409531] env[61243]: DEBUG oslo_concurrency.lockutils [None req-485b5121-df60-4547-80c3-b439fdd08d9a tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Lock "6f33b303-1ecd-478e-8fa3-33a5f9a5af50" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.711s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.465644] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338885, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.115495} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.466534] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.467459] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8efb79ce-7103-4751-8330-083ea39021d1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.496036] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 1dc8ad6a-3a00-47c6-8985-481e7f1363cd/1dc8ad6a-3a00-47c6-8985-481e7f1363cd.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.498075] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e62988d9-3d45-4c7e-bee9-8845768bac39 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.519944] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 846.519944] env[61243]: value = "task-1338887" [ 846.519944] env[61243]: _type = "Task" [ 846.519944] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.529445] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338887, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.586912] env[61243]: DEBUG nova.network.neutron [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Successfully created port: e21d7ddc-a0d3-4026-a937-f392d5e20fa4 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 846.615393] env[61243]: DEBUG nova.compute.manager [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 846.684429] env[61243]: DEBUG oslo_concurrency.lockutils [req-6af5fb47-e9ba-4afc-ada5-c771e6328011 req-d5b61a72-dd4c-4098-8613-1588e37ac17d service nova] Releasing lock "refresh_cache-ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.787980] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338886, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.883242] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338883, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.910448] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Releasing lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.910790] env[61243]: DEBUG nova.compute.manager [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Instance network_info: |[{"id": "817c2755-c3a0-4de3-9427-322d92ba1d30", "address": "fa:16:3e:5c:f9:55", "network": {"id": "f6dda51c-55ed-4b80-b54d-ade24626e5fd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1249632116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93625b58c8674dc0980e3d9fee62ab18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap817c2755-c3", "ovs_interfaceid": "817c2755-c3a0-4de3-9427-322d92ba1d30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 846.911120] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f50975-8b5f-7a45-157d-3c2621b521f5, 'name': SearchDatastore_Task, 'duration_secs': 0.031811} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.913956] env[61243]: DEBUG oslo_concurrency.lockutils [req-d89132e3-2c52-4474-abec-631981a52ccc req-6e0830d0-a5c1-451d-adf4-9cb0ed363372 service nova] Acquired lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.914037] env[61243]: DEBUG nova.network.neutron [req-d89132e3-2c52-4474-abec-631981a52ccc req-6e0830d0-a5c1-451d-adf4-9cb0ed363372 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Refreshing network info cache for port 817c2755-c3a0-4de3-9427-322d92ba1d30 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 846.915316] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:f9:55', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f8442aa5-73db-4599-8564-b98a6ea26b9b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '817c2755-c3a0-4de3-9427-322d92ba1d30', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 846.922711] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Creating folder: Project (93625b58c8674dc0980e3d9fee62ab18). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 846.923133] env[61243]: DEBUG nova.compute.manager [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 846.926100] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.926591] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 846.926742] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.926922] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.927147] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.928579] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5bfcc2e1-3413-406f-8636-9219f52fba5d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.930458] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ebc75f48-ea2e-4790-ac63-02833b9e8f7d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.945034] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Created folder: Project (93625b58c8674dc0980e3d9fee62ab18) in parent group-v285636. [ 846.945034] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Creating folder: Instances. Parent ref: group-v285695. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 846.945310] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.945480] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 846.946280] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a56e9f3-3670-4dea-b439-8ee217cc67ec {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.948580] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da4ca852-3f23-4e20-b125-8ce56887d2e0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.955501] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 846.955501] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d18967-b25f-f91c-693b-b5115d0b1018" [ 846.955501] env[61243]: _type = "Task" [ 846.955501] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.961909] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Created folder: Instances in parent group-v285695. [ 846.962138] env[61243]: DEBUG oslo.service.loopingcall [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.962454] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 846.963014] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-386b2e6d-dd5b-46ef-82b9-676f943c2d22 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.985797] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d18967-b25f-f91c-693b-b5115d0b1018, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.990080] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 846.990080] env[61243]: value = "task-1338890" [ 846.990080] env[61243]: _type = "Task" [ 846.990080] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.999012] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338890, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.001085] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e60fceb6-c8a1-4b6f-bbf9-eccdc398cf6b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.011823] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60676c49-1f3b-47a5-9fba-faa6760b9e89 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.016469] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a70a96c8-a86c-4947-9abc-c472ef55554e tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Acquiring lock "interface-6f33b303-1ecd-478e-8fa3-33a5f9a5af50-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.016795] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a70a96c8-a86c-4947-9abc-c472ef55554e tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Lock "interface-6f33b303-1ecd-478e-8fa3-33a5f9a5af50-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.017152] env[61243]: DEBUG nova.objects.instance [None req-a70a96c8-a86c-4947-9abc-c472ef55554e tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Lazy-loading 'flavor' on Instance uuid 6f33b303-1ecd-478e-8fa3-33a5f9a5af50 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.050936] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e762fdf-e4ec-469d-8406-a9d43bd01e54 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.058645] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338887, 'name': ReconfigVM_Task, 'duration_secs': 0.293031} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.058645] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 1dc8ad6a-3a00-47c6-8985-481e7f1363cd/1dc8ad6a-3a00-47c6-8985-481e7f1363cd.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.059126] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1b0655a6-ab8f-48a9-9ac8-5048ffc58bad {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.065362] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610995ba-43d5-45cd-b395-75f1ffb4a506 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.074162] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 847.074162] env[61243]: value = "task-1338891" [ 847.074162] env[61243]: _type = "Task" [ 847.074162] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.089039] env[61243]: DEBUG nova.compute.provider_tree [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.096028] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338891, 'name': Rename_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.288329] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338886, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.887858} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.288656] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 96315d9e-4eda-4e3a-af0d-bdc52ab181e3/96315d9e-4eda-4e3a-af0d-bdc52ab181e3.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 847.288838] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 847.289819] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-159ccc74-027b-4b9c-8caa-8d9c59e4fab6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.296941] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 847.296941] env[61243]: value = "task-1338892" [ 847.296941] env[61243]: _type = "Task" [ 847.296941] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.308575] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338892, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.382827] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338883, 'name': PowerOnVM_Task, 'duration_secs': 1.836952} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.383153] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 847.383448] env[61243]: INFO nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Took 9.63 seconds to spawn the instance on the hypervisor. [ 847.383691] env[61243]: DEBUG nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 847.384797] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72db433-f1f7-41b4-907a-be864c9c81bc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.451828] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.470028] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d18967-b25f-f91c-693b-b5115d0b1018, 'name': SearchDatastore_Task, 'duration_secs': 0.021356} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.471025] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8b39af2-bb18-4c5e-bde6-8e81f32fa54e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.476322] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 847.476322] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d8c3bf-9b8d-657d-0733-9245537b2084" [ 847.476322] env[61243]: _type = "Task" [ 847.476322] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.484424] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d8c3bf-9b8d-657d-0733-9245537b2084, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.498863] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338890, 'name': CreateVM_Task, 'duration_secs': 0.42723} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.499682] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 847.499980] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.500405] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.500504] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 847.500768] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b101debd-e6f2-44ef-a90f-d917c964ef62 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.505186] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Waiting for the task: (returnval){ [ 847.505186] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52993eac-2ede-1252-6c48-c2bd6d8d79c9" [ 847.505186] env[61243]: _type = "Task" [ 847.505186] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.514906] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52993eac-2ede-1252-6c48-c2bd6d8d79c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.521730] env[61243]: DEBUG nova.objects.instance [None req-a70a96c8-a86c-4947-9abc-c472ef55554e tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Lazy-loading 'pci_requests' on Instance uuid 6f33b303-1ecd-478e-8fa3-33a5f9a5af50 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.586594] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338891, 'name': Rename_Task, 'duration_secs': 0.191041} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.586908] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 847.587244] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ba85f7ba-dbba-4c4a-a79e-92912ece038d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.592622] env[61243]: DEBUG nova.scheduler.client.report [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 847.598219] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 847.598219] env[61243]: value = "task-1338893" [ 847.598219] env[61243]: _type = "Task" [ 847.598219] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.610467] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338893, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.630222] env[61243]: DEBUG nova.compute.manager [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 847.672022] env[61243]: DEBUG nova.virt.hardware [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.672022] env[61243]: DEBUG nova.virt.hardware [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.672022] env[61243]: DEBUG nova.virt.hardware [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.672341] env[61243]: DEBUG nova.virt.hardware [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.672341] env[61243]: DEBUG nova.virt.hardware [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.672545] env[61243]: DEBUG nova.virt.hardware [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.672820] env[61243]: DEBUG nova.virt.hardware [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.673021] env[61243]: DEBUG nova.virt.hardware [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.673261] env[61243]: DEBUG nova.virt.hardware [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.673470] env[61243]: DEBUG nova.virt.hardware [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.673684] env[61243]: DEBUG nova.virt.hardware [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.674735] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e138c80-17a2-4dce-aa33-8ba807f44a29 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.684317] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d767b869-3558-4327-80db-768767f6ab1f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.806503] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338892, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06591} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.806779] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.807669] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-553c3b6e-e7ea-41cd-b18b-b3b503b9eab5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.828212] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 96315d9e-4eda-4e3a-af0d-bdc52ab181e3/96315d9e-4eda-4e3a-af0d-bdc52ab181e3.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.829701] env[61243]: DEBUG nova.network.neutron [req-d89132e3-2c52-4474-abec-631981a52ccc req-6e0830d0-a5c1-451d-adf4-9cb0ed363372 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Updated VIF entry in instance network info cache for port 817c2755-c3a0-4de3-9427-322d92ba1d30. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 847.829701] env[61243]: DEBUG nova.network.neutron [req-d89132e3-2c52-4474-abec-631981a52ccc req-6e0830d0-a5c1-451d-adf4-9cb0ed363372 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Updating instance_info_cache with network_info: [{"id": "817c2755-c3a0-4de3-9427-322d92ba1d30", "address": "fa:16:3e:5c:f9:55", "network": {"id": "f6dda51c-55ed-4b80-b54d-ade24626e5fd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1249632116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93625b58c8674dc0980e3d9fee62ab18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap817c2755-c3", "ovs_interfaceid": "817c2755-c3a0-4de3-9427-322d92ba1d30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.830939] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b5c687a-5b5a-4f6b-ae05-30f8fa948c22 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.852482] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 847.852482] env[61243]: value = "task-1338894" [ 847.852482] env[61243]: _type = "Task" [ 847.852482] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.860854] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338894, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.904287] env[61243]: INFO nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Took 24.45 seconds to build instance. [ 847.988248] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d8c3bf-9b8d-657d-0733-9245537b2084, 'name': SearchDatastore_Task, 'duration_secs': 0.011852} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.988764] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.988935] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5/ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 847.990073] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f5ab4aea-32c0-45d1-a7d7-8851831a2c88 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.996188] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 847.996188] env[61243]: value = "task-1338895" [ 847.996188] env[61243]: _type = "Task" [ 847.996188] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.005311] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338895, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.014506] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52993eac-2ede-1252-6c48-c2bd6d8d79c9, 'name': SearchDatastore_Task, 'duration_secs': 0.009545} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.015804] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.015804] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 848.015804] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.015804] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.016017] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 848.016049] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fba6ba7e-6ce2-4815-ba3b-94c4fadc4f48 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.024629] env[61243]: DEBUG nova.objects.base [None req-a70a96c8-a86c-4947-9abc-c472ef55554e tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Object Instance<6f33b303-1ecd-478e-8fa3-33a5f9a5af50> lazy-loaded attributes: flavor,pci_requests {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 848.024629] env[61243]: DEBUG nova.network.neutron [None req-a70a96c8-a86c-4947-9abc-c472ef55554e tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 848.028563] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 848.028750] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 848.029500] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-191cf554-b287-4617-81ec-03512993f0e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.034532] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Waiting for the task: (returnval){ [ 848.034532] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5286c88a-3693-6fd9-d92d-5c7837423705" [ 848.034532] env[61243]: _type = "Task" [ 848.034532] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.049943] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5286c88a-3693-6fd9-d92d-5c7837423705, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.099453] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.099982] env[61243]: DEBUG nova.compute.manager [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 848.103218] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.748s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.107017] env[61243]: INFO nova.compute.claims [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.118102] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338893, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.173382] env[61243]: DEBUG nova.network.neutron [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Successfully updated port: e21d7ddc-a0d3-4026-a937-f392d5e20fa4 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 848.177321] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a70a96c8-a86c-4947-9abc-c472ef55554e tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Lock "interface-6f33b303-1ecd-478e-8fa3-33a5f9a5af50-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.160s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.241605] env[61243]: DEBUG nova.compute.manager [req-09587adf-7012-4387-a218-f3a8629d58aa req-7e82fe99-8077-48d7-a3f5-38caac60ac00 service nova] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Received event network-vif-plugged-e21d7ddc-a0d3-4026-a937-f392d5e20fa4 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.241911] env[61243]: DEBUG oslo_concurrency.lockutils [req-09587adf-7012-4387-a218-f3a8629d58aa req-7e82fe99-8077-48d7-a3f5-38caac60ac00 service nova] Acquiring lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.242132] env[61243]: DEBUG oslo_concurrency.lockutils [req-09587adf-7012-4387-a218-f3a8629d58aa req-7e82fe99-8077-48d7-a3f5-38caac60ac00 service nova] Lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.242254] env[61243]: DEBUG oslo_concurrency.lockutils [req-09587adf-7012-4387-a218-f3a8629d58aa req-7e82fe99-8077-48d7-a3f5-38caac60ac00 service nova] Lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.242460] env[61243]: DEBUG nova.compute.manager [req-09587adf-7012-4387-a218-f3a8629d58aa req-7e82fe99-8077-48d7-a3f5-38caac60ac00 service nova] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] No waiting events found dispatching network-vif-plugged-e21d7ddc-a0d3-4026-a937-f392d5e20fa4 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 848.242638] env[61243]: WARNING nova.compute.manager [req-09587adf-7012-4387-a218-f3a8629d58aa req-7e82fe99-8077-48d7-a3f5-38caac60ac00 service nova] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Received unexpected event network-vif-plugged-e21d7ddc-a0d3-4026-a937-f392d5e20fa4 for instance with vm_state building and task_state spawning. [ 848.348461] env[61243]: DEBUG oslo_concurrency.lockutils [req-d89132e3-2c52-4474-abec-631981a52ccc req-6e0830d0-a5c1-451d-adf4-9cb0ed363372 service nova] Releasing lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.362961] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338894, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.406980] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "9a66fd05-1880-423a-8f60-4f11f1c75ab3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.527s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.506577] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338895, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.544295] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5286c88a-3693-6fd9-d92d-5c7837423705, 'name': SearchDatastore_Task, 'duration_secs': 0.018685} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.545217] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f770432-4928-46d7-ad6d-fea584b1c78f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.551161] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Waiting for the task: (returnval){ [ 848.551161] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5235c749-7099-f52b-f121-b903b21fad24" [ 848.551161] env[61243]: _type = "Task" [ 848.551161] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.559432] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5235c749-7099-f52b-f121-b903b21fad24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.614057] env[61243]: DEBUG nova.compute.utils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 848.617455] env[61243]: DEBUG oslo_vmware.api [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338893, 'name': PowerOnVM_Task, 'duration_secs': 0.97685} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.618124] env[61243]: DEBUG nova.compute.manager [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 848.618315] env[61243]: DEBUG nova.network.neutron [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 848.620176] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 848.620430] env[61243]: INFO nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Took 8.38 seconds to spawn the instance on the hypervisor. [ 848.620621] env[61243]: DEBUG nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 848.621474] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46bf7d53-9afe-4526-8471-6cc5d350c7cc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.668921] env[61243]: DEBUG nova.policy [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '876a16a17dd34418b20be920e53779f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2465299214be4d5f877f340489215d6d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 848.675807] env[61243]: DEBUG oslo_concurrency.lockutils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.675998] env[61243]: DEBUG oslo_concurrency.lockutils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.676204] env[61243]: DEBUG nova.network.neutron [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 848.867405] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338894, 'name': ReconfigVM_Task, 'duration_secs': 0.813589} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.867405] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 96315d9e-4eda-4e3a-af0d-bdc52ab181e3/96315d9e-4eda-4e3a-af0d-bdc52ab181e3.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.867405] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-48fa1800-0beb-4df8-830f-98ebd4bf45e1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.875068] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 848.875068] env[61243]: value = "task-1338896" [ 848.875068] env[61243]: _type = "Task" [ 848.875068] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.882978] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338896, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.912971] env[61243]: DEBUG nova.compute.manager [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 848.981419] env[61243]: DEBUG nova.network.neutron [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Successfully created port: 970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.010174] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338895, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.971805} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.010732] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5/ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 849.010732] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 849.011011] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-af18b2db-6ca4-441f-b1bc-fb23e2bf4da4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.018049] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 849.018049] env[61243]: value = "task-1338897" [ 849.018049] env[61243]: _type = "Task" [ 849.018049] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.027139] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338897, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.066401] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5235c749-7099-f52b-f121-b903b21fad24, 'name': SearchDatastore_Task, 'duration_secs': 0.030392} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.066719] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.067027] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 36db1ed5-846f-4ad6-8cee-38b73ff00321/36db1ed5-846f-4ad6-8cee-38b73ff00321.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 849.067321] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dbe5bd31-0699-4fc5-aac0-5a2318bfa794 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.073542] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Waiting for the task: (returnval){ [ 849.073542] env[61243]: value = "task-1338898" [ 849.073542] env[61243]: _type = "Task" [ 849.073542] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.082453] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1338898, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.121924] env[61243]: DEBUG nova.compute.manager [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 849.146843] env[61243]: INFO nova.compute.manager [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Took 25.19 seconds to build instance. [ 849.231164] env[61243]: DEBUG nova.network.neutron [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 849.385457] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338896, 'name': Rename_Task, 'duration_secs': 0.213465} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.385859] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 849.385983] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-98ad05fe-d620-4bae-a533-b17e460c6b82 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.392655] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 849.392655] env[61243]: value = "task-1338899" [ 849.392655] env[61243]: _type = "Task" [ 849.392655] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.401383] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338899, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.418337] env[61243]: DEBUG nova.network.neutron [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating instance_info_cache with network_info: [{"id": "e21d7ddc-a0d3-4026-a937-f392d5e20fa4", "address": "fa:16:3e:56:77:58", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape21d7ddc-a0", "ovs_interfaceid": "e21d7ddc-a0d3-4026-a937-f392d5e20fa4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.432854] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.444954] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c75947e-0232-4a88-9883-82d44997f12f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.453172] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e9e2244-6fc3-42b2-9dd3-8a7fb7e01c87 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.484225] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cabe4af-a37b-403a-b234-c1f24fb64032 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.492033] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf495bef-5f82-475c-b4f3-d4857f7c2f47 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.508423] env[61243]: DEBUG nova.compute.provider_tree [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.528036] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338897, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068346} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.528187] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 849.528903] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17016935-e729-4694-8d36-ca152412bed5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.554301] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5/ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 849.554703] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f561b8c-aac8-46ef-99a2-be63dda9e8b0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.580137] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 849.580137] env[61243]: value = "task-1338900" [ 849.580137] env[61243]: _type = "Task" [ 849.580137] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.586871] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1338898, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.593917] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338900, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.653655] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d8214fc-9718-4491-8b94-c22af4817107 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "1dc8ad6a-3a00-47c6-8985-481e7f1363cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.736s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.881456] env[61243]: DEBUG oslo_concurrency.lockutils [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "9a66fd05-1880-423a-8f60-4f11f1c75ab3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.881904] env[61243]: DEBUG oslo_concurrency.lockutils [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "9a66fd05-1880-423a-8f60-4f11f1c75ab3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.882170] env[61243]: DEBUG oslo_concurrency.lockutils [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "9a66fd05-1880-423a-8f60-4f11f1c75ab3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.882572] env[61243]: DEBUG oslo_concurrency.lockutils [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "9a66fd05-1880-423a-8f60-4f11f1c75ab3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.882922] env[61243]: DEBUG oslo_concurrency.lockutils [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "9a66fd05-1880-423a-8f60-4f11f1c75ab3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.889221] env[61243]: INFO nova.compute.manager [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Terminating instance [ 849.892194] env[61243]: DEBUG nova.compute.manager [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 849.892194] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 849.893382] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2215cd9-95c7-4f3f-8533-69257c01ce3d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.905706] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 849.909358] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80ca6acd-8cb6-4d32-aade-7fa371752ab9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.911748] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338899, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.919217] env[61243]: DEBUG oslo_vmware.api [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 849.919217] env[61243]: value = "task-1338901" [ 849.919217] env[61243]: _type = "Task" [ 849.919217] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.923637] env[61243]: DEBUG oslo_concurrency.lockutils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.924034] env[61243]: DEBUG nova.compute.manager [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Instance network_info: |[{"id": "e21d7ddc-a0d3-4026-a937-f392d5e20fa4", "address": "fa:16:3e:56:77:58", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape21d7ddc-a0", "ovs_interfaceid": "e21d7ddc-a0d3-4026-a937-f392d5e20fa4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 849.924478] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:56:77:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd7b5f1ef-d4b9-4ec3-b047-17e4cb349d25', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e21d7ddc-a0d3-4026-a937-f392d5e20fa4', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 849.932733] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Creating folder: Project (66e2ba61b3354ef1b496232264d27600). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.933424] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1f96f3b8-a09e-4a09-b1a0-7147570054d2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.938239] env[61243]: DEBUG oslo_vmware.api [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338901, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.945626] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Created folder: Project (66e2ba61b3354ef1b496232264d27600) in parent group-v285636. [ 849.945819] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Creating folder: Instances. Parent ref: group-v285698. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.946152] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1004d908-fc53-41d9-8b8c-a51402ec95b1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.955741] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Created folder: Instances in parent group-v285698. [ 849.956134] env[61243]: DEBUG oslo.service.loopingcall [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.956371] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 849.956627] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-211718dc-3bdb-41ad-aeb3-e658559d1e42 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.975845] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.975845] env[61243]: value = "task-1338904" [ 849.975845] env[61243]: _type = "Task" [ 849.975845] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.986512] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338904, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.012201] env[61243]: DEBUG nova.scheduler.client.report [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 850.087036] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1338898, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.890369} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.087036] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 36db1ed5-846f-4ad6-8cee-38b73ff00321/36db1ed5-846f-4ad6-8cee-38b73ff00321.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 850.087222] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 850.087357] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1e229278-5427-4faf-9c28-1c514f3afc10 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.094231] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338900, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.095531] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Waiting for the task: (returnval){ [ 850.095531] env[61243]: value = "task-1338905" [ 850.095531] env[61243]: _type = "Task" [ 850.095531] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.103907] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1338905, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.137900] env[61243]: DEBUG nova.compute.manager [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 850.157785] env[61243]: DEBUG nova.compute.manager [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 850.167236] env[61243]: DEBUG nova.virt.hardware [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.167494] env[61243]: DEBUG nova.virt.hardware [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.167662] env[61243]: DEBUG nova.virt.hardware [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.167851] env[61243]: DEBUG nova.virt.hardware [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.168012] env[61243]: DEBUG nova.virt.hardware [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.168174] env[61243]: DEBUG nova.virt.hardware [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.168465] env[61243]: DEBUG nova.virt.hardware [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.168557] env[61243]: DEBUG nova.virt.hardware [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.168749] env[61243]: DEBUG nova.virt.hardware [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.168955] env[61243]: DEBUG nova.virt.hardware [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.169315] env[61243]: DEBUG nova.virt.hardware [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.170055] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46345ce8-394c-45ab-adfc-c02c93bc90df {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.179054] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812fe64e-1419-4c94-bee5-a54ec2d4a65e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.183860] env[61243]: DEBUG oslo_concurrency.lockutils [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "1dc8ad6a-3a00-47c6-8985-481e7f1363cd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.183860] env[61243]: DEBUG oslo_concurrency.lockutils [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "1dc8ad6a-3a00-47c6-8985-481e7f1363cd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.184924] env[61243]: DEBUG oslo_concurrency.lockutils [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "1dc8ad6a-3a00-47c6-8985-481e7f1363cd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.184924] env[61243]: DEBUG oslo_concurrency.lockutils [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "1dc8ad6a-3a00-47c6-8985-481e7f1363cd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.184924] env[61243]: DEBUG oslo_concurrency.lockutils [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "1dc8ad6a-3a00-47c6-8985-481e7f1363cd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.187421] env[61243]: INFO nova.compute.manager [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Terminating instance [ 850.197424] env[61243]: DEBUG nova.compute.manager [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 850.197642] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 850.198458] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb864f86-4aca-4213-9668-4aad8b5a81f5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.206459] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 850.206743] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7746318b-adca-4052-8f8c-ef8b043ed472 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.215187] env[61243]: DEBUG oslo_vmware.api [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 850.215187] env[61243]: value = "task-1338906" [ 850.215187] env[61243]: _type = "Task" [ 850.215187] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.227353] env[61243]: DEBUG oslo_vmware.api [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338906, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.262554] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Acquiring lock "6f33b303-1ecd-478e-8fa3-33a5f9a5af50" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.262835] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Lock "6f33b303-1ecd-478e-8fa3-33a5f9a5af50" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.263066] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Acquiring lock "6f33b303-1ecd-478e-8fa3-33a5f9a5af50-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.263265] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Lock "6f33b303-1ecd-478e-8fa3-33a5f9a5af50-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.263457] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Lock "6f33b303-1ecd-478e-8fa3-33a5f9a5af50-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.265929] env[61243]: INFO nova.compute.manager [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Terminating instance [ 850.269680] env[61243]: DEBUG nova.compute.manager [req-87626a62-9f77-4e09-be0f-2fe5284de39f req-54933f63-81cb-434e-b339-68637b2b43a1 service nova] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Received event network-changed-e21d7ddc-a0d3-4026-a937-f392d5e20fa4 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.269933] env[61243]: DEBUG nova.compute.manager [req-87626a62-9f77-4e09-be0f-2fe5284de39f req-54933f63-81cb-434e-b339-68637b2b43a1 service nova] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Refreshing instance network info cache due to event network-changed-e21d7ddc-a0d3-4026-a937-f392d5e20fa4. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 850.270142] env[61243]: DEBUG oslo_concurrency.lockutils [req-87626a62-9f77-4e09-be0f-2fe5284de39f req-54933f63-81cb-434e-b339-68637b2b43a1 service nova] Acquiring lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.270294] env[61243]: DEBUG oslo_concurrency.lockutils [req-87626a62-9f77-4e09-be0f-2fe5284de39f req-54933f63-81cb-434e-b339-68637b2b43a1 service nova] Acquired lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.270529] env[61243]: DEBUG nova.network.neutron [req-87626a62-9f77-4e09-be0f-2fe5284de39f req-54933f63-81cb-434e-b339-68637b2b43a1 service nova] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Refreshing network info cache for port e21d7ddc-a0d3-4026-a937-f392d5e20fa4 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 850.273022] env[61243]: DEBUG nova.compute.manager [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 850.273022] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 850.273388] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36ebae4a-ffd5-45d0-9374-2af2386934e5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.281639] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 850.282374] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8b4cf31-0790-432b-b847-7e0c584e50bb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.290346] env[61243]: DEBUG oslo_vmware.api [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Waiting for the task: (returnval){ [ 850.290346] env[61243]: value = "task-1338907" [ 850.290346] env[61243]: _type = "Task" [ 850.290346] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.303321] env[61243]: DEBUG oslo_vmware.api [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338907, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.408531] env[61243]: DEBUG oslo_vmware.api [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338899, 'name': PowerOnVM_Task, 'duration_secs': 0.661864} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.408917] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 850.409217] env[61243]: DEBUG nova.compute.manager [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 850.410188] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d45c11-5b95-469f-8943-2a45ec0d4f2d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.430419] env[61243]: DEBUG oslo_vmware.api [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338901, 'name': PowerOffVM_Task, 'duration_secs': 0.338276} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.431131] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 850.431131] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 850.431389] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-03fc4b8b-790c-41aa-bb30-ae015bbee91c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.485092] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338904, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.518730] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.415s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.519340] env[61243]: DEBUG nova.compute.manager [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 850.522102] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 11.036s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.522295] env[61243]: DEBUG nova.objects.instance [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61243) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 850.594030] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338900, 'name': ReconfigVM_Task, 'duration_secs': 0.740042} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.594954] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Reconfigured VM instance instance-00000043 to attach disk [datastore2] ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5/ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.595979] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-85a32a9d-a6c9-42f3-a2c6-4c63f77e9f2e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.606571] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1338905, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.221174} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.609085] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 850.609085] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 850.609085] env[61243]: value = "task-1338909" [ 850.609085] env[61243]: _type = "Task" [ 850.609085] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.609085] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39da7041-03ef-4d61-92ac-7a9c0b16a201 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.620970] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338909, 'name': Rename_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.644042] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 36db1ed5-846f-4ad6-8cee-38b73ff00321/36db1ed5-846f-4ad6-8cee-38b73ff00321.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 850.644042] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d636d809-69b4-4ff8-9ef8-ac144b9e3464 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.659917] env[61243]: DEBUG nova.network.neutron [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Successfully updated port: 970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 850.669026] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Waiting for the task: (returnval){ [ 850.669026] env[61243]: value = "task-1338910" [ 850.669026] env[61243]: _type = "Task" [ 850.669026] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.678887] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1338910, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.680481] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.725136] env[61243]: DEBUG oslo_vmware.api [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338906, 'name': PowerOffVM_Task, 'duration_secs': 0.4217} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.725541] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 850.725900] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 850.726839] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-69595524-9a36-4c1d-a217-041a981063c8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.800653] env[61243]: DEBUG oslo_vmware.api [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338907, 'name': PowerOffVM_Task, 'duration_secs': 0.264388} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.800958] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 850.801154] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 850.801421] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aaf11fd6-346a-4b5c-8e53-38fa88dace86 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.928435] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.994925] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338904, 'name': CreateVM_Task, 'duration_secs': 0.721464} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.995195] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 850.996183] env[61243]: DEBUG nova.network.neutron [req-87626a62-9f77-4e09-be0f-2fe5284de39f req-54933f63-81cb-434e-b339-68637b2b43a1 service nova] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updated VIF entry in instance network info cache for port e21d7ddc-a0d3-4026-a937-f392d5e20fa4. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 850.996533] env[61243]: DEBUG nova.network.neutron [req-87626a62-9f77-4e09-be0f-2fe5284de39f req-54933f63-81cb-434e-b339-68637b2b43a1 service nova] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating instance_info_cache with network_info: [{"id": "e21d7ddc-a0d3-4026-a937-f392d5e20fa4", "address": "fa:16:3e:56:77:58", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape21d7ddc-a0", "ovs_interfaceid": "e21d7ddc-a0d3-4026-a937-f392d5e20fa4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.998320] env[61243]: DEBUG oslo_concurrency.lockutils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.998512] env[61243]: DEBUG oslo_concurrency.lockutils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.998855] env[61243]: DEBUG oslo_concurrency.lockutils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 850.999343] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a835fad-11d9-45c8-96dd-d164c690ea78 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.004296] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 851.004296] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5284bc87-bcff-f09a-fc4d-f3e839f6d23a" [ 851.004296] env[61243]: _type = "Task" [ 851.004296] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.014662] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5284bc87-bcff-f09a-fc4d-f3e839f6d23a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.017944] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 851.017944] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 851.017944] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Deleting the datastore file [datastore2] 9a66fd05-1880-423a-8f60-4f11f1c75ab3 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 851.017944] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-17dcb1bb-be7d-491d-b05d-7e0a6aa92e3b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.021686] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 851.021884] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 851.022072] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Deleting the datastore file [datastore2] 6f33b303-1ecd-478e-8fa3-33a5f9a5af50 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 851.022331] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 851.022511] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 851.022675] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Deleting the datastore file [datastore2] 1dc8ad6a-3a00-47c6-8985-481e7f1363cd {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 851.023238] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5832f8c1-c373-41f7-812e-8e7fa1e49b1e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.024875] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e3234beb-8511-4a1b-82dc-e48d40c88061 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.033522] env[61243]: DEBUG nova.compute.utils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 851.034933] env[61243]: DEBUG oslo_vmware.api [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 851.034933] env[61243]: value = "task-1338913" [ 851.034933] env[61243]: _type = "Task" [ 851.034933] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.036688] env[61243]: DEBUG nova.compute.manager [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 851.036688] env[61243]: DEBUG nova.network.neutron [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 851.044631] env[61243]: DEBUG oslo_vmware.api [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Waiting for the task: (returnval){ [ 851.044631] env[61243]: value = "task-1338914" [ 851.044631] env[61243]: _type = "Task" [ 851.044631] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.044780] env[61243]: DEBUG oslo_vmware.api [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 851.044780] env[61243]: value = "task-1338915" [ 851.044780] env[61243]: _type = "Task" [ 851.044780] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.054995] env[61243]: DEBUG oslo_vmware.api [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338913, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.060377] env[61243]: DEBUG oslo_vmware.api [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338914, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.063136] env[61243]: DEBUG oslo_vmware.api [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338915, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.084205] env[61243]: DEBUG nova.policy [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62fa373ddccf48e084524b6f2b5987f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fbbffd1228af409b818182e8c20ee02a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 851.121227] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338909, 'name': Rename_Task, 'duration_secs': 0.382372} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.121522] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 851.121828] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-55af89a2-df09-4d25-ab71-915502e0dbb5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.127850] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 851.127850] env[61243]: value = "task-1338916" [ 851.127850] env[61243]: _type = "Task" [ 851.127850] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.136364] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338916, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.164392] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.164614] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.164838] env[61243]: DEBUG nova.network.neutron [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 851.179463] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1338910, 'name': ReconfigVM_Task, 'duration_secs': 0.284848} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.179765] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 36db1ed5-846f-4ad6-8cee-38b73ff00321/36db1ed5-846f-4ad6-8cee-38b73ff00321.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 851.180482] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69ad0a20-13f5-4d3b-aa41-044dfce87953 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.187368] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Waiting for the task: (returnval){ [ 851.187368] env[61243]: value = "task-1338917" [ 851.187368] env[61243]: _type = "Task" [ 851.187368] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.196842] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1338917, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.353175] env[61243]: DEBUG nova.network.neutron [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Successfully created port: c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 851.500884] env[61243]: DEBUG oslo_concurrency.lockutils [req-87626a62-9f77-4e09-be0f-2fe5284de39f req-54933f63-81cb-434e-b339-68637b2b43a1 service nova] Releasing lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.515675] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5284bc87-bcff-f09a-fc4d-f3e839f6d23a, 'name': SearchDatastore_Task, 'duration_secs': 0.014785} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.516053] env[61243]: DEBUG oslo_concurrency.lockutils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.516269] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 851.516579] env[61243]: DEBUG oslo_concurrency.lockutils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.516778] env[61243]: DEBUG oslo_concurrency.lockutils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.516925] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 851.517219] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c6dc405e-cb0d-4438-b399-549973add33a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.528661] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 851.528851] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 851.529596] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef149162-cb4e-4ca8-983d-09e4ae37e8fb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.534865] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 851.534865] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5206b872-7116-d8d6-9fcf-a2e65beee65a" [ 851.534865] env[61243]: _type = "Task" [ 851.534865] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.538908] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a9398c51-c343-47db-809d-abaa722b2b67 tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.540439] env[61243]: DEBUG nova.compute.manager [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 851.543189] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.319s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.544412] env[61243]: INFO nova.compute.claims [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.559690] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5206b872-7116-d8d6-9fcf-a2e65beee65a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.565260] env[61243]: DEBUG oslo_vmware.api [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338913, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.163152} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.571377] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 851.571545] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 851.571951] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 851.571951] env[61243]: INFO nova.compute.manager [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Took 1.68 seconds to destroy the instance on the hypervisor. [ 851.572266] env[61243]: DEBUG oslo.service.loopingcall [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.572534] env[61243]: DEBUG oslo_vmware.api [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1338915, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178525} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.572822] env[61243]: DEBUG oslo_vmware.api [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Task: {'id': task-1338914, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164653} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.572973] env[61243]: DEBUG nova.compute.manager [-] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 851.575880] env[61243]: DEBUG nova.network.neutron [-] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 851.577756] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 851.577957] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 851.578238] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 851.578481] env[61243]: INFO nova.compute.manager [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Took 1.38 seconds to destroy the instance on the hypervisor. [ 851.579361] env[61243]: DEBUG oslo.service.loopingcall [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.579361] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 851.579361] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 851.579523] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 851.579594] env[61243]: INFO nova.compute.manager [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Took 1.31 seconds to destroy the instance on the hypervisor. [ 851.580166] env[61243]: DEBUG oslo.service.loopingcall [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.580994] env[61243]: DEBUG nova.compute.manager [-] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 851.581113] env[61243]: DEBUG nova.network.neutron [-] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 851.582793] env[61243]: DEBUG nova.compute.manager [-] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 851.582925] env[61243]: DEBUG nova.network.neutron [-] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 851.638809] env[61243]: DEBUG oslo_vmware.api [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338916, 'name': PowerOnVM_Task, 'duration_secs': 0.496374} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.639017] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 851.639235] env[61243]: INFO nova.compute.manager [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Took 8.96 seconds to spawn the instance on the hypervisor. [ 851.639635] env[61243]: DEBUG nova.compute.manager [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.640643] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc27fce-ed44-43a5-ad40-923c2846cb3e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.698374] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1338917, 'name': Rename_Task, 'duration_secs': 0.145396} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.699360] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 851.699663] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52195950-cdf3-4068-ae3f-5c2d059aacd5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.707639] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Waiting for the task: (returnval){ [ 851.707639] env[61243]: value = "task-1338918" [ 851.707639] env[61243]: _type = "Task" [ 851.707639] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.724278] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1338918, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.747088] env[61243]: DEBUG nova.network.neutron [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 851.959769] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "96315d9e-4eda-4e3a-af0d-bdc52ab181e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.960079] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "96315d9e-4eda-4e3a-af0d-bdc52ab181e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.960326] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "96315d9e-4eda-4e3a-af0d-bdc52ab181e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.960528] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "96315d9e-4eda-4e3a-af0d-bdc52ab181e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.960727] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "96315d9e-4eda-4e3a-af0d-bdc52ab181e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.963012] env[61243]: INFO nova.compute.manager [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Terminating instance [ 851.965094] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "refresh_cache-96315d9e-4eda-4e3a-af0d-bdc52ab181e3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.971153] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquired lock "refresh_cache-96315d9e-4eda-4e3a-af0d-bdc52ab181e3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.971153] env[61243]: DEBUG nova.network.neutron [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 851.972052] env[61243]: DEBUG nova.network.neutron [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Updating instance_info_cache with network_info: [{"id": "970e8509-164d-4ddf-8f0c-795e92883f6b", "address": "fa:16:3e:00:af:7d", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap970e8509-16", "ovs_interfaceid": "970e8509-164d-4ddf-8f0c-795e92883f6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.046485] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5206b872-7116-d8d6-9fcf-a2e65beee65a, 'name': SearchDatastore_Task, 'duration_secs': 0.029165} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.047488] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc93e1c7-5532-4d80-909b-a5533837280f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.063360] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 852.063360] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52bb8d17-e309-5cea-b8a0-742e1db84678" [ 852.063360] env[61243]: _type = "Task" [ 852.063360] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.078980] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52bb8d17-e309-5cea-b8a0-742e1db84678, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.159385] env[61243]: INFO nova.compute.manager [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Took 27.69 seconds to build instance. [ 852.217701] env[61243]: DEBUG oslo_vmware.api [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1338918, 'name': PowerOnVM_Task, 'duration_secs': 0.483747} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.217951] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 852.218505] env[61243]: INFO nova.compute.manager [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Took 7.01 seconds to spawn the instance on the hypervisor. [ 852.218741] env[61243]: DEBUG nova.compute.manager [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.219535] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1ddc2cd-9a0e-405f-a64f-43605fc4a195 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.475723] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Releasing lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.480142] env[61243]: DEBUG nova.compute.manager [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Instance network_info: |[{"id": "970e8509-164d-4ddf-8f0c-795e92883f6b", "address": "fa:16:3e:00:af:7d", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap970e8509-16", "ovs_interfaceid": "970e8509-164d-4ddf-8f0c-795e92883f6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 852.480673] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:af:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d413776-9a8c-4afd-856f-10dbb062ca95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '970e8509-164d-4ddf-8f0c-795e92883f6b', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 852.492591] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Creating folder: Project (2465299214be4d5f877f340489215d6d). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 852.494449] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-95017860-4b89-4c95-aab7-7d9a6ab0962c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.508045] env[61243]: DEBUG nova.network.neutron [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.516326] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Created folder: Project (2465299214be4d5f877f340489215d6d) in parent group-v285636. [ 852.516562] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Creating folder: Instances. Parent ref: group-v285701. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 852.518018] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a04411e8-640e-48b4-96d1-7d0c6c24f75f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.529197] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Created folder: Instances in parent group-v285701. [ 852.529440] env[61243]: DEBUG oslo.service.loopingcall [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.529630] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 852.529841] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-099b3615-95f4-42e3-afc9-69a81bfe2e7b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.548192] env[61243]: DEBUG nova.compute.manager [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Received event network-vif-plugged-970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.548472] env[61243]: DEBUG oslo_concurrency.lockutils [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] Acquiring lock "c1a0d242-4eae-4c03-8341-840b41341f17-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.548786] env[61243]: DEBUG oslo_concurrency.lockutils [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] Lock "c1a0d242-4eae-4c03-8341-840b41341f17-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.548860] env[61243]: DEBUG oslo_concurrency.lockutils [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] Lock "c1a0d242-4eae-4c03-8341-840b41341f17-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.549203] env[61243]: DEBUG nova.compute.manager [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] No waiting events found dispatching network-vif-plugged-970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 852.549302] env[61243]: WARNING nova.compute.manager [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Received unexpected event network-vif-plugged-970e8509-164d-4ddf-8f0c-795e92883f6b for instance with vm_state building and task_state spawning. [ 852.549392] env[61243]: DEBUG nova.compute.manager [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Received event network-changed-970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.549549] env[61243]: DEBUG nova.compute.manager [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Refreshing instance network info cache due to event network-changed-970e8509-164d-4ddf-8f0c-795e92883f6b. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 852.549737] env[61243]: DEBUG oslo_concurrency.lockutils [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] Acquiring lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.549871] env[61243]: DEBUG oslo_concurrency.lockutils [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] Acquired lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.550061] env[61243]: DEBUG nova.network.neutron [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Refreshing network info cache for port 970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 852.559096] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 852.559096] env[61243]: value = "task-1338921" [ 852.559096] env[61243]: _type = "Task" [ 852.559096] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.564116] env[61243]: DEBUG nova.compute.manager [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 852.574281] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338921, 'name': CreateVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.575229] env[61243]: DEBUG nova.network.neutron [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.581815] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52bb8d17-e309-5cea-b8a0-742e1db84678, 'name': SearchDatastore_Task, 'duration_secs': 0.031162} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.582122] env[61243]: DEBUG oslo_concurrency.lockutils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.582519] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 316572e0-c007-42cb-aaf0-3a8cfcaf24aa/316572e0-c007-42cb-aaf0-3a8cfcaf24aa.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 852.582850] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a05f802c-76d3-4646-9fb6-949007e00743 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.592955] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 852.592955] env[61243]: value = "task-1338922" [ 852.592955] env[61243]: _type = "Task" [ 852.592955] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.595287] env[61243]: DEBUG nova.virt.hardware [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 852.595532] env[61243]: DEBUG nova.virt.hardware [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 852.595697] env[61243]: DEBUG nova.virt.hardware [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 852.595884] env[61243]: DEBUG nova.virt.hardware [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 852.596048] env[61243]: DEBUG nova.virt.hardware [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 852.596208] env[61243]: DEBUG nova.virt.hardware [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 852.596438] env[61243]: DEBUG nova.virt.hardware [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 852.596621] env[61243]: DEBUG nova.virt.hardware [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 852.596808] env[61243]: DEBUG nova.virt.hardware [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 852.596979] env[61243]: DEBUG nova.virt.hardware [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 852.597226] env[61243]: DEBUG nova.virt.hardware [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 852.598344] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44156aec-864e-4be5-8dde-056d46278ab7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.613327] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f395f3b-3a23-452d-ade2-13303a7b1368 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.617745] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1338922, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.664776] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a4a7624a-d9df-4a1e-a133-2015e782a48d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.727s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.687307] env[61243]: DEBUG nova.network.neutron [-] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.698971] env[61243]: DEBUG nova.network.neutron [-] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.703140] env[61243]: DEBUG nova.network.neutron [-] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.745045] env[61243]: INFO nova.compute.manager [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Took 26.25 seconds to build instance. [ 852.978487] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a3caf88-69b1-45f5-bb42-49760d9dc661 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.987264] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94f49dec-854d-432f-ac00-fb834233bdb4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.024725] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.024994] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.025237] env[61243]: INFO nova.compute.manager [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Shelving [ 853.030595] env[61243]: DEBUG nova.network.neutron [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Successfully updated port: c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 853.030595] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-018af85b-c1a1-4443-bd92-2c32bbce2eca {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.039263] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd277e17-dda8-4e00-b1d3-fb938a8831ad {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.056956] env[61243]: DEBUG nova.compute.provider_tree [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 853.070062] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338921, 'name': CreateVM_Task, 'duration_secs': 0.393985} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.070253] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 853.071102] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.071207] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.071454] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 853.071996] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb63a764-8765-4869-bc2a-5c1c4bf48bcf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.076950] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 853.076950] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529bc586-f12d-7646-cbde-b0a4a83f04b9" [ 853.076950] env[61243]: _type = "Task" [ 853.076950] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.077392] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Releasing lock "refresh_cache-96315d9e-4eda-4e3a-af0d-bdc52ab181e3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.077866] env[61243]: DEBUG nova.compute.manager [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 853.078086] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 853.079176] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-648b8d4a-a27c-43a6-b7f1-19ae92f2a142 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.090013] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 853.095676] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-173fcbe6-199d-4a66-bf07-a7388f2c22e6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.101018] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529bc586-f12d-7646-cbde-b0a4a83f04b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.102861] env[61243]: DEBUG oslo_vmware.api [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 853.102861] env[61243]: value = "task-1338923" [ 853.102861] env[61243]: _type = "Task" [ 853.102861] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.110710] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1338922, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.116980] env[61243]: DEBUG oslo_vmware.api [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338923, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.167563] env[61243]: DEBUG nova.compute.manager [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 853.189823] env[61243]: INFO nova.compute.manager [-] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Took 1.62 seconds to deallocate network for instance. [ 853.206830] env[61243]: INFO nova.compute.manager [-] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Took 1.63 seconds to deallocate network for instance. [ 853.207200] env[61243]: INFO nova.compute.manager [-] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Took 1.62 seconds to deallocate network for instance. [ 853.248026] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9ddce16a-d123-43bb-9079-13a34c3a7f5d tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Lock "36db1ed5-846f-4ad6-8cee-38b73ff00321" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.102s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.323021] env[61243]: DEBUG nova.network.neutron [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Updated VIF entry in instance network info cache for port 970e8509-164d-4ddf-8f0c-795e92883f6b. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 853.323432] env[61243]: DEBUG nova.network.neutron [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Updating instance_info_cache with network_info: [{"id": "970e8509-164d-4ddf-8f0c-795e92883f6b", "address": "fa:16:3e:00:af:7d", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap970e8509-16", "ovs_interfaceid": "970e8509-164d-4ddf-8f0c-795e92883f6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.535320] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Acquiring lock "refresh_cache-f73a4d39-5478-4135-9be6-e59f3e29788d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.535320] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Acquired lock "refresh_cache-f73a4d39-5478-4135-9be6-e59f3e29788d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.535320] env[61243]: DEBUG nova.network.neutron [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.536857] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 853.537138] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-300949e5-bfb9-45aa-9182-51b94b6367fa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.544561] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 853.544561] env[61243]: value = "task-1338924" [ 853.544561] env[61243]: _type = "Task" [ 853.544561] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.556015] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338924, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.584526] env[61243]: ERROR nova.scheduler.client.report [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [req-732ef02a-7016-44b2-9dda-d6e02beefc54] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-732ef02a-7016-44b2-9dda-d6e02beefc54"}]} [ 853.591913] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529bc586-f12d-7646-cbde-b0a4a83f04b9, 'name': SearchDatastore_Task, 'duration_secs': 0.053617} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.593098] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.593098] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 853.593098] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.593098] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.593328] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 853.593360] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4a8af9ee-f807-4cd3-9bcc-04709f067ac5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.605396] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 853.605396] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 853.606532] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c5792fb-f9ef-47d7-9bf7-587f22401c48 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.609200] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1338922, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56305} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.613783] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 316572e0-c007-42cb-aaf0-3a8cfcaf24aa/316572e0-c007-42cb-aaf0-3a8cfcaf24aa.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 853.614088] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 853.614845] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed835f22-2f4d-4b26-862f-955a22b36a94 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.618815] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 853.618815] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]526dfdeb-88fd-cdb1-318b-35af7231b8f6" [ 853.618815] env[61243]: _type = "Task" [ 853.618815] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.619721] env[61243]: DEBUG nova.scheduler.client.report [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 853.624284] env[61243]: DEBUG oslo_vmware.api [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338923, 'name': PowerOffVM_Task, 'duration_secs': 0.1215} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.633311] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 853.633503] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 853.633894] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 853.633894] env[61243]: value = "task-1338925" [ 853.633894] env[61243]: _type = "Task" [ 853.633894] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.634376] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8776193-0a2f-421b-923d-a312aacb0297 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.643956] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]526dfdeb-88fd-cdb1-318b-35af7231b8f6, 'name': SearchDatastore_Task, 'duration_secs': 0.011358} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.644909] env[61243]: DEBUG nova.scheduler.client.report [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 853.645135] env[61243]: DEBUG nova.compute.provider_tree [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 853.649268] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3fa8adb-46c8-4895-b45c-e0c9e47630f1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.652884] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1338925, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.656061] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 853.656061] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52eae45f-0ee7-6442-25bb-93995a6ef5ba" [ 853.656061] env[61243]: _type = "Task" [ 853.656061] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.660810] env[61243]: DEBUG nova.scheduler.client.report [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 853.666493] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52eae45f-0ee7-6442-25bb-93995a6ef5ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.667696] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 853.667914] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 853.668181] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Deleting the datastore file [datastore1] 96315d9e-4eda-4e3a-af0d-bdc52ab181e3 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 853.668459] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-17837752-4126-4750-890b-d54483fa999a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.675436] env[61243]: DEBUG oslo_vmware.api [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 853.675436] env[61243]: value = "task-1338927" [ 853.675436] env[61243]: _type = "Task" [ 853.675436] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.686119] env[61243]: DEBUG oslo_vmware.api [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338927, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.691361] env[61243]: DEBUG oslo_concurrency.lockutils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.693975] env[61243]: DEBUG nova.scheduler.client.report [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 853.697169] env[61243]: DEBUG oslo_concurrency.lockutils [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.718009] env[61243]: DEBUG oslo_concurrency.lockutils [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.730912] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.826341] env[61243]: DEBUG oslo_concurrency.lockutils [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] Releasing lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.826645] env[61243]: DEBUG nova.compute.manager [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Received event network-vif-deleted-f80550c9-ffa4-4f9f-a0c1-7be2562defa3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.826836] env[61243]: INFO nova.compute.manager [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Neutron deleted interface f80550c9-ffa4-4f9f-a0c1-7be2562defa3; detaching it from the instance and deleting it from the info cache [ 853.827034] env[61243]: DEBUG nova.network.neutron [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.976199] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40d84f2d-6e8f-451a-8701-fcbe3304ccbe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.984657] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ad17ae-2810-40d6-b81a-bf663d851f88 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.019721] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7504cb6c-27f8-4514-b4e2-2424007c09c6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.027474] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e743d8c3-413e-4688-92d8-49294950783b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.045676] env[61243]: DEBUG nova.compute.provider_tree [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 854.052988] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338924, 'name': PowerOffVM_Task, 'duration_secs': 0.270798} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.053334] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 854.054147] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d17f49-e657-41be-a578-11d8d01921e3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.072784] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d2fba43-8686-4a0b-842c-5405d04f37e9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.086351] env[61243]: DEBUG nova.network.neutron [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 854.119294] env[61243]: DEBUG nova.compute.manager [req-2b4f82e9-de22-4367-931f-61195061c47d req-48ac9db2-fbd8-402c-b158-45c3d54930ba service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Received event network-changed-817c2755-c3a0-4de3-9427-322d92ba1d30 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.119294] env[61243]: DEBUG nova.compute.manager [req-2b4f82e9-de22-4367-931f-61195061c47d req-48ac9db2-fbd8-402c-b158-45c3d54930ba service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Refreshing instance network info cache due to event network-changed-817c2755-c3a0-4de3-9427-322d92ba1d30. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 854.119294] env[61243]: DEBUG oslo_concurrency.lockutils [req-2b4f82e9-de22-4367-931f-61195061c47d req-48ac9db2-fbd8-402c-b158-45c3d54930ba service nova] Acquiring lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.119294] env[61243]: DEBUG oslo_concurrency.lockutils [req-2b4f82e9-de22-4367-931f-61195061c47d req-48ac9db2-fbd8-402c-b158-45c3d54930ba service nova] Acquired lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.119294] env[61243]: DEBUG nova.network.neutron [req-2b4f82e9-de22-4367-931f-61195061c47d req-48ac9db2-fbd8-402c-b158-45c3d54930ba service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Refreshing network info cache for port 817c2755-c3a0-4de3-9427-322d92ba1d30 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 854.149055] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1338925, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066191} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.149055] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.149610] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f65cff-8f76-4eb1-9068-ab64fd1a08cb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.176677] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 316572e0-c007-42cb-aaf0-3a8cfcaf24aa/316572e0-c007-42cb-aaf0-3a8cfcaf24aa.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.180532] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de10115b-5469-412e-ba6b-62e843ff0b8b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.206070] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52eae45f-0ee7-6442-25bb-93995a6ef5ba, 'name': SearchDatastore_Task, 'duration_secs': 0.013321} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.210041] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.210396] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] c1a0d242-4eae-4c03-8341-840b41341f17/c1a0d242-4eae-4c03-8341-840b41341f17.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 854.210780] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 854.210780] env[61243]: value = "task-1338928" [ 854.210780] env[61243]: _type = "Task" [ 854.210780] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.211107] env[61243]: DEBUG oslo_vmware.api [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1338927, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179458} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.211367] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-405e280c-4ae8-4697-9221-de76d16546fb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.213349] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 854.213611] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 854.213855] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 854.214184] env[61243]: INFO nova.compute.manager [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Took 1.14 seconds to destroy the instance on the hypervisor. [ 854.214449] env[61243]: DEBUG oslo.service.loopingcall [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.217779] env[61243]: DEBUG nova.compute.manager [-] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 854.217869] env[61243]: DEBUG nova.network.neutron [-] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 854.226759] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1338928, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.226759] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 854.226759] env[61243]: value = "task-1338929" [ 854.226759] env[61243]: _type = "Task" [ 854.226759] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.234173] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1338929, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.241304] env[61243]: DEBUG nova.network.neutron [-] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 854.295464] env[61243]: DEBUG nova.network.neutron [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Updating instance_info_cache with network_info: [{"id": "c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3", "address": "fa:16:3e:5f:ad:1c", "network": {"id": "3a682e54-55e6-44a3-aaa0-827525351b06", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1993848006-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbbffd1228af409b818182e8c20ee02a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc640e6df-91", "ovs_interfaceid": "c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.329930] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2529efc1-0f74-4d80-a679-3a957ee35d03 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.338627] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d52d309-0be8-43e7-af62-81d61936f0aa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.366667] env[61243]: DEBUG nova.compute.manager [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Detach interface failed, port_id=f80550c9-ffa4-4f9f-a0c1-7be2562defa3, reason: Instance 9a66fd05-1880-423a-8f60-4f11f1c75ab3 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 854.367150] env[61243]: DEBUG nova.compute.manager [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Received event network-vif-deleted-ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.367150] env[61243]: INFO nova.compute.manager [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Neutron deleted interface ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6; detaching it from the instance and deleting it from the info cache [ 854.367313] env[61243]: DEBUG nova.network.neutron [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.566387] env[61243]: DEBUG nova.compute.manager [req-75846efb-03c0-42e7-ac5b-31f9d0139262 req-898a299d-d6bb-487d-8229-afa9000bd7d6 service nova] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Received event network-vif-plugged-c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.566697] env[61243]: DEBUG oslo_concurrency.lockutils [req-75846efb-03c0-42e7-ac5b-31f9d0139262 req-898a299d-d6bb-487d-8229-afa9000bd7d6 service nova] Acquiring lock "f73a4d39-5478-4135-9be6-e59f3e29788d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.566903] env[61243]: DEBUG oslo_concurrency.lockutils [req-75846efb-03c0-42e7-ac5b-31f9d0139262 req-898a299d-d6bb-487d-8229-afa9000bd7d6 service nova] Lock "f73a4d39-5478-4135-9be6-e59f3e29788d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.567133] env[61243]: DEBUG oslo_concurrency.lockutils [req-75846efb-03c0-42e7-ac5b-31f9d0139262 req-898a299d-d6bb-487d-8229-afa9000bd7d6 service nova] Lock "f73a4d39-5478-4135-9be6-e59f3e29788d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.567367] env[61243]: DEBUG nova.compute.manager [req-75846efb-03c0-42e7-ac5b-31f9d0139262 req-898a299d-d6bb-487d-8229-afa9000bd7d6 service nova] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] No waiting events found dispatching network-vif-plugged-c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 854.567557] env[61243]: WARNING nova.compute.manager [req-75846efb-03c0-42e7-ac5b-31f9d0139262 req-898a299d-d6bb-487d-8229-afa9000bd7d6 service nova] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Received unexpected event network-vif-plugged-c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3 for instance with vm_state building and task_state spawning. [ 854.567721] env[61243]: DEBUG nova.compute.manager [req-75846efb-03c0-42e7-ac5b-31f9d0139262 req-898a299d-d6bb-487d-8229-afa9000bd7d6 service nova] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Received event network-changed-c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.567880] env[61243]: DEBUG nova.compute.manager [req-75846efb-03c0-42e7-ac5b-31f9d0139262 req-898a299d-d6bb-487d-8229-afa9000bd7d6 service nova] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Refreshing instance network info cache due to event network-changed-c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 854.568065] env[61243]: DEBUG oslo_concurrency.lockutils [req-75846efb-03c0-42e7-ac5b-31f9d0139262 req-898a299d-d6bb-487d-8229-afa9000bd7d6 service nova] Acquiring lock "refresh_cache-f73a4d39-5478-4135-9be6-e59f3e29788d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.578669] env[61243]: DEBUG nova.scheduler.client.report [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 90 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 854.579129] env[61243]: DEBUG nova.compute.provider_tree [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 90 to 91 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 854.579266] env[61243]: DEBUG nova.compute.provider_tree [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 854.583884] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Creating Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 854.584713] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-344b6a2b-7066-4347-b918-3d775dc3c94c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.594099] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 854.594099] env[61243]: value = "task-1338930" [ 854.594099] env[61243]: _type = "Task" [ 854.594099] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.605405] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338930, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.724413] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1338928, 'name': ReconfigVM_Task, 'duration_secs': 0.237218} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.727026] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 316572e0-c007-42cb-aaf0-3a8cfcaf24aa/316572e0-c007-42cb-aaf0-3a8cfcaf24aa.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 854.727764] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8216560d-dd8c-419c-bcb6-89799642c0d2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.738085] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1338929, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.740020] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] c1a0d242-4eae-4c03-8341-840b41341f17/c1a0d242-4eae-4c03-8341-840b41341f17.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 854.740020] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.740212] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 854.740212] env[61243]: value = "task-1338931" [ 854.740212] env[61243]: _type = "Task" [ 854.740212] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.740345] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0510b546-cfd1-4246-9adf-9f43c459640d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.742460] env[61243]: DEBUG nova.network.neutron [-] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.752189] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1338931, 'name': Rename_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.753614] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 854.753614] env[61243]: value = "task-1338932" [ 854.753614] env[61243]: _type = "Task" [ 854.753614] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.766522] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1338932, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.798613] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Releasing lock "refresh_cache-f73a4d39-5478-4135-9be6-e59f3e29788d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.799029] env[61243]: DEBUG nova.compute.manager [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Instance network_info: |[{"id": "c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3", "address": "fa:16:3e:5f:ad:1c", "network": {"id": "3a682e54-55e6-44a3-aaa0-827525351b06", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1993848006-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbbffd1228af409b818182e8c20ee02a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc640e6df-91", "ovs_interfaceid": "c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 854.799380] env[61243]: DEBUG oslo_concurrency.lockutils [req-75846efb-03c0-42e7-ac5b-31f9d0139262 req-898a299d-d6bb-487d-8229-afa9000bd7d6 service nova] Acquired lock "refresh_cache-f73a4d39-5478-4135-9be6-e59f3e29788d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.799588] env[61243]: DEBUG nova.network.neutron [req-75846efb-03c0-42e7-ac5b-31f9d0139262 req-898a299d-d6bb-487d-8229-afa9000bd7d6 service nova] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Refreshing network info cache for port c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 854.800891] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5f:ad:1c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '184687d6-125a-4b58-bb5b-fdb404088eda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.808590] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Creating folder: Project (fbbffd1228af409b818182e8c20ee02a). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 854.810025] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f54453c-cda7-4fe8-9150-87662d4818b6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.820468] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Created folder: Project (fbbffd1228af409b818182e8c20ee02a) in parent group-v285636. [ 854.821296] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Creating folder: Instances. Parent ref: group-v285704. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 854.821296] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-174397a5-c944-4d45-a1fa-19c2be07205f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.829632] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Created folder: Instances in parent group-v285704. [ 854.829867] env[61243]: DEBUG oslo.service.loopingcall [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.830075] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 854.830387] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6d180f5-ed9c-468e-9d14-1a7acbe7bfb1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.850364] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 854.850364] env[61243]: value = "task-1338935" [ 854.850364] env[61243]: _type = "Task" [ 854.850364] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.858157] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338935, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.871532] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6d4b711-c727-4d95-a780-0e186d2ec108 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.881711] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ff4015-2220-4ac9-b514-e990d33b7a4c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.894514] env[61243]: DEBUG nova.network.neutron [req-2b4f82e9-de22-4367-931f-61195061c47d req-48ac9db2-fbd8-402c-b158-45c3d54930ba service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Updated VIF entry in instance network info cache for port 817c2755-c3a0-4de3-9427-322d92ba1d30. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 854.894947] env[61243]: DEBUG nova.network.neutron [req-2b4f82e9-de22-4367-931f-61195061c47d req-48ac9db2-fbd8-402c-b158-45c3d54930ba service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Updating instance_info_cache with network_info: [{"id": "817c2755-c3a0-4de3-9427-322d92ba1d30", "address": "fa:16:3e:5c:f9:55", "network": {"id": "f6dda51c-55ed-4b80-b54d-ade24626e5fd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1249632116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93625b58c8674dc0980e3d9fee62ab18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap817c2755-c3", "ovs_interfaceid": "817c2755-c3a0-4de3-9427-322d92ba1d30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.917389] env[61243]: DEBUG nova.compute.manager [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Detach interface failed, port_id=ebf3bb9d-c47b-4c8f-94f3-d60d58e568a6, reason: Instance 1dc8ad6a-3a00-47c6-8985-481e7f1363cd could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 854.917756] env[61243]: DEBUG nova.compute.manager [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Received event network-vif-deleted-f2472adb-26ae-4651-9108-d8286beeb98b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.918058] env[61243]: INFO nova.compute.manager [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Neutron deleted interface f2472adb-26ae-4651-9108-d8286beeb98b; detaching it from the instance and deleting it from the info cache [ 854.918292] env[61243]: DEBUG nova.network.neutron [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.085606] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.542s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.086145] env[61243]: DEBUG nova.compute.manager [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 855.088876] env[61243]: DEBUG oslo_concurrency.lockutils [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.935s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.106203] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338930, 'name': CreateSnapshot_Task, 'duration_secs': 0.499884} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.106619] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Created Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 855.107600] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c68c8386-9c82-4c30-8a3e-58493b2021a5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.249164] env[61243]: INFO nova.compute.manager [-] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Took 1.03 seconds to deallocate network for instance. [ 855.260577] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1338931, 'name': Rename_Task, 'duration_secs': 0.131839} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.264197] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 855.264749] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-145748d3-9467-48ea-af4a-7754f8abcef7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.271523] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1338932, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063779} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.274844] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 855.275268] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 855.275268] env[61243]: value = "task-1338936" [ 855.275268] env[61243]: _type = "Task" [ 855.275268] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.276365] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74261b44-1beb-4852-813b-a83f5c3f9971 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.302437] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] c1a0d242-4eae-4c03-8341-840b41341f17/c1a0d242-4eae-4c03-8341-840b41341f17.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.308982] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd0d0e2a-2b99-438e-aef1-7f825322a1b6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.324075] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1338936, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.332881] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 855.332881] env[61243]: value = "task-1338937" [ 855.332881] env[61243]: _type = "Task" [ 855.332881] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.344396] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1338937, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.359736] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338935, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.397571] env[61243]: DEBUG oslo_concurrency.lockutils [req-2b4f82e9-de22-4367-931f-61195061c47d req-48ac9db2-fbd8-402c-b158-45c3d54930ba service nova] Releasing lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.421183] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f3ad86cc-cbd0-41f3-9340-bad95f2f1934 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.429426] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf82e93-919f-460b-9cf0-1523d39fdbc3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.463686] env[61243]: DEBUG nova.compute.manager [req-f678886d-b8fd-482b-ad30-5c74a2d5a679 req-9c7f9327-fa6c-442c-8ea2-de9763240200 service nova] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Detach interface failed, port_id=f2472adb-26ae-4651-9108-d8286beeb98b, reason: Instance 6f33b303-1ecd-478e-8fa3-33a5f9a5af50 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 855.467510] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee00af5-b183-4bd5-bd51-935c8ff628a8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.474610] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae0c571-43ea-4b4f-b977-f24db271960a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.506465] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701146ea-3241-4135-9b91-58192aefbf60 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.514038] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1fbc998-ac4a-40dd-9ac8-a7ec113697b8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.528460] env[61243]: DEBUG nova.compute.provider_tree [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.579212] env[61243]: DEBUG nova.network.neutron [req-75846efb-03c0-42e7-ac5b-31f9d0139262 req-898a299d-d6bb-487d-8229-afa9000bd7d6 service nova] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Updated VIF entry in instance network info cache for port c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 855.579212] env[61243]: DEBUG nova.network.neutron [req-75846efb-03c0-42e7-ac5b-31f9d0139262 req-898a299d-d6bb-487d-8229-afa9000bd7d6 service nova] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Updating instance_info_cache with network_info: [{"id": "c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3", "address": "fa:16:3e:5f:ad:1c", "network": {"id": "3a682e54-55e6-44a3-aaa0-827525351b06", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1993848006-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbbffd1228af409b818182e8c20ee02a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc640e6df-91", "ovs_interfaceid": "c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.594047] env[61243]: DEBUG nova.compute.utils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 855.595659] env[61243]: DEBUG nova.compute.manager [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 855.595659] env[61243]: DEBUG nova.network.neutron [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 855.630733] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Creating linked-clone VM from snapshot {{(pid=61243) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 855.631382] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-824fa3da-29b6-4f10-bdfa-38f39e1ac06f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.636718] env[61243]: DEBUG nova.policy [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4dc60c433f984e70917f48bab6c02b1c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '399030c8863346ea97f6da669ff3868b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 855.640520] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 855.640520] env[61243]: value = "task-1338938" [ 855.640520] env[61243]: _type = "Task" [ 855.640520] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.648603] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338938, 'name': CloneVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.763155] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.789039] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1338936, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.843141] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1338937, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.860515] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338935, 'name': CreateVM_Task, 'duration_secs': 0.695397} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.860811] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 855.861946] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.862291] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.862738] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.863119] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da920013-aa8c-40c6-9df3-79c8c3884548 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.869556] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Waiting for the task: (returnval){ [ 855.869556] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52caa661-2d03-ca5d-e188-bcc7e35005ca" [ 855.869556] env[61243]: _type = "Task" [ 855.869556] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.884774] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52caa661-2d03-ca5d-e188-bcc7e35005ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.912354] env[61243]: DEBUG nova.network.neutron [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Successfully created port: 1aa2d7ff-eac6-472b-b7a7-57c542d3f111 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 856.032870] env[61243]: DEBUG nova.scheduler.client.report [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.080811] env[61243]: DEBUG oslo_concurrency.lockutils [req-75846efb-03c0-42e7-ac5b-31f9d0139262 req-898a299d-d6bb-487d-8229-afa9000bd7d6 service nova] Releasing lock "refresh_cache-f73a4d39-5478-4135-9be6-e59f3e29788d" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.099068] env[61243]: DEBUG nova.compute.manager [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 856.152243] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338938, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.291618] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1338936, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.343313] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1338937, 'name': ReconfigVM_Task, 'duration_secs': 0.603572} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.343684] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Reconfigured VM instance instance-00000046 to attach disk [datastore1] c1a0d242-4eae-4c03-8341-840b41341f17/c1a0d242-4eae-4c03-8341-840b41341f17.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 856.344431] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-adcfe5b2-a258-442c-8b06-591ec7395767 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.351197] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 856.351197] env[61243]: value = "task-1338939" [ 856.351197] env[61243]: _type = "Task" [ 856.351197] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.359643] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1338939, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.380316] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52caa661-2d03-ca5d-e188-bcc7e35005ca, 'name': SearchDatastore_Task, 'duration_secs': 0.020388} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.380657] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.380921] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 856.381172] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.381327] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.381516] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 856.381797] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f1339363-f1df-4847-b51d-dc59e45d86cd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.390329] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 856.390596] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 856.391391] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5b1ad4e-946e-4be4-8ff0-9ca907e7cf57 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.396875] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Waiting for the task: (returnval){ [ 856.396875] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522072c4-29da-9760-405b-3db35131b784" [ 856.396875] env[61243]: _type = "Task" [ 856.396875] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.404792] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522072c4-29da-9760-405b-3db35131b784, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.538070] env[61243]: DEBUG oslo_concurrency.lockutils [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.449s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.538333] env[61243]: INFO nova.compute.manager [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Successfully reverted task state from image_uploading on failure for instance. [ 856.541176] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.977s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.542651] env[61243]: INFO nova.compute.claims [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 856.551182] env[61243]: ERROR oslo_messaging.rpc.server [None req-000375ad-95ce-4ca6-8802-f65d073b158a tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Exception during message handling: oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-285664' has already been deleted or has not been completely created [ 856.551182] env[61243]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 856.551182] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 856.551182] env[61243]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 856.551182] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 856.551182] env[61243]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 856.551182] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 856.551182] env[61243]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 856.551182] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 856.551182] env[61243]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 856.551182] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 856.551182] env[61243]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 856.551182] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 856.551182] env[61243]: ERROR oslo_messaging.rpc.server raise self.value [ 856.551182] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server raise self.value [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server raise self.value [ 856.552023] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 233, in decorated_function [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server raise self.value [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 230, in decorated_function [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server return function(self, context, image_id, instance, [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 4442, in snapshot_instance [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server self._snapshot_instance(context, image_id, instance, [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 4475, in _snapshot_instance [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server self.driver.snapshot(context, instance, image_id, [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 571, in snapshot [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server self._vmops.snapshot(context, instance, image_id, update_task_state) [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1033, in snapshot [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server self._delete_vm_snapshot(instance, vm_ref, snapshot_ref) [ 856.552835] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/decorator.py", line 232, in fun [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server return caller(func, *(extras + args), **kw) [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 124, in retry_if_task_in_progress [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server f(*args, **kwargs) [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 937, in _delete_vm_snapshot [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server self._session._wait_for_task(delete_snapshot_task) [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server return self.wait_for_task(task_ref) [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 398, in wait_for_task [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server return evt.wait() [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server result = hub.switch() [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server return self.greenlet.switch() [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server self.f(*self.args, **self.kw) [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 448, in _poll_task [ 856.553733] env[61243]: ERROR oslo_messaging.rpc.server raise exceptions.translate_fault(task_info.error) [ 856.554581] env[61243]: ERROR oslo_messaging.rpc.server oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-285664' has already been deleted or has not been completely created [ 856.554581] env[61243]: ERROR oslo_messaging.rpc.server [ 856.654538] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338938, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.789729] env[61243]: DEBUG oslo_vmware.api [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1338936, 'name': PowerOnVM_Task, 'duration_secs': 1.152939} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.790014] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 856.790240] env[61243]: INFO nova.compute.manager [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Took 9.16 seconds to spawn the instance on the hypervisor. [ 856.790422] env[61243]: DEBUG nova.compute.manager [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.791186] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a21665-957d-417d-ab04-4a98cb7901f1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.860504] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1338939, 'name': Rename_Task, 'duration_secs': 0.336557} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.860808] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.861071] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-99210efa-58ad-4ba5-a370-e6d8ed33f115 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.866848] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 856.866848] env[61243]: value = "task-1338940" [ 856.866848] env[61243]: _type = "Task" [ 856.866848] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.874293] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1338940, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.909489] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522072c4-29da-9760-405b-3db35131b784, 'name': SearchDatastore_Task, 'duration_secs': 0.011605} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.910303] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b980f7b2-99b3-4edb-b185-236ebdd80876 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.915767] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Waiting for the task: (returnval){ [ 856.915767] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a7dba0-1cc1-2999-d5fb-8eb5048e74b0" [ 856.915767] env[61243]: _type = "Task" [ 856.915767] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.923535] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a7dba0-1cc1-2999-d5fb-8eb5048e74b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.107605] env[61243]: DEBUG nova.compute.manager [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 857.139718] env[61243]: DEBUG nova.virt.hardware [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 857.139977] env[61243]: DEBUG nova.virt.hardware [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 857.140225] env[61243]: DEBUG nova.virt.hardware [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 857.140431] env[61243]: DEBUG nova.virt.hardware [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 857.140580] env[61243]: DEBUG nova.virt.hardware [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 857.140734] env[61243]: DEBUG nova.virt.hardware [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 857.140945] env[61243]: DEBUG nova.virt.hardware [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 857.141122] env[61243]: DEBUG nova.virt.hardware [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 857.141298] env[61243]: DEBUG nova.virt.hardware [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 857.141463] env[61243]: DEBUG nova.virt.hardware [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 857.141639] env[61243]: DEBUG nova.virt.hardware [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 857.142526] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08dc061a-8992-4d20-84ac-476ef20231bc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.155208] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338938, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.156418] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7d8040-899f-468f-8c45-5ffe6f5117de {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.310328] env[61243]: DEBUG nova.compute.manager [req-1549dc9f-785d-49f6-a183-12581ba700f6 req-d19c84f0-b984-4aaa-92ff-cb807c105982 service nova] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Received event network-vif-plugged-1aa2d7ff-eac6-472b-b7a7-57c542d3f111 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.310726] env[61243]: DEBUG oslo_concurrency.lockutils [req-1549dc9f-785d-49f6-a183-12581ba700f6 req-d19c84f0-b984-4aaa-92ff-cb807c105982 service nova] Acquiring lock "5daccb48-b9bf-423a-ab1a-5aa26d8bec87-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.311046] env[61243]: DEBUG oslo_concurrency.lockutils [req-1549dc9f-785d-49f6-a183-12581ba700f6 req-d19c84f0-b984-4aaa-92ff-cb807c105982 service nova] Lock "5daccb48-b9bf-423a-ab1a-5aa26d8bec87-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.311046] env[61243]: DEBUG oslo_concurrency.lockutils [req-1549dc9f-785d-49f6-a183-12581ba700f6 req-d19c84f0-b984-4aaa-92ff-cb807c105982 service nova] Lock "5daccb48-b9bf-423a-ab1a-5aa26d8bec87-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.311363] env[61243]: DEBUG nova.compute.manager [req-1549dc9f-785d-49f6-a183-12581ba700f6 req-d19c84f0-b984-4aaa-92ff-cb807c105982 service nova] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] No waiting events found dispatching network-vif-plugged-1aa2d7ff-eac6-472b-b7a7-57c542d3f111 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 857.311363] env[61243]: WARNING nova.compute.manager [req-1549dc9f-785d-49f6-a183-12581ba700f6 req-d19c84f0-b984-4aaa-92ff-cb807c105982 service nova] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Received unexpected event network-vif-plugged-1aa2d7ff-eac6-472b-b7a7-57c542d3f111 for instance with vm_state building and task_state spawning. [ 857.311829] env[61243]: INFO nova.compute.manager [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Took 29.25 seconds to build instance. [ 857.377044] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1338940, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.406830] env[61243]: DEBUG nova.network.neutron [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Successfully updated port: 1aa2d7ff-eac6-472b-b7a7-57c542d3f111 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.429083] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a7dba0-1cc1-2999-d5fb-8eb5048e74b0, 'name': SearchDatastore_Task, 'duration_secs': 0.00958} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.430738] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.431209] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] f73a4d39-5478-4135-9be6-e59f3e29788d/f73a4d39-5478-4135-9be6-e59f3e29788d.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 857.431892] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d6803d0-72f2-4a8e-887e-bcaa6a7fda76 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.441178] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Waiting for the task: (returnval){ [ 857.441178] env[61243]: value = "task-1338941" [ 857.441178] env[61243]: _type = "Task" [ 857.441178] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.449074] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': task-1338941, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.657707] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338938, 'name': CloneVM_Task, 'duration_secs': 1.879632} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.658068] env[61243]: INFO nova.virt.vmwareapi.vmops [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Created linked-clone VM from snapshot [ 857.658864] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e381ae86-e0a4-466a-87f8-570c41d6bac3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.673457] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Uploading image daf672ed-2d6b-4ef8-8911-3a353c295ac4 {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 857.706429] env[61243]: DEBUG oslo_vmware.rw_handles [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 857.706429] env[61243]: value = "vm-285708" [ 857.706429] env[61243]: _type = "VirtualMachine" [ 857.706429] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 857.706694] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-989a8387-1ac3-4e59-83c2-f5810c755da3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.717625] env[61243]: DEBUG oslo_vmware.rw_handles [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lease: (returnval){ [ 857.717625] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52df14fb-4eb5-380a-7937-fd0eb03b8a7b" [ 857.717625] env[61243]: _type = "HttpNfcLease" [ 857.717625] env[61243]: } obtained for exporting VM: (result){ [ 857.717625] env[61243]: value = "vm-285708" [ 857.717625] env[61243]: _type = "VirtualMachine" [ 857.717625] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 857.718335] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the lease: (returnval){ [ 857.718335] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52df14fb-4eb5-380a-7937-fd0eb03b8a7b" [ 857.718335] env[61243]: _type = "HttpNfcLease" [ 857.718335] env[61243]: } to be ready. {{(pid=61243) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 857.733279] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 857.733279] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52df14fb-4eb5-380a-7937-fd0eb03b8a7b" [ 857.733279] env[61243]: _type = "HttpNfcLease" [ 857.733279] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 857.745343] env[61243]: DEBUG nova.compute.manager [req-33f05a7b-f106-43c5-83ca-ff3e237d392c req-530fc4cf-32e3-4189-8ee7-048faa0db130 service nova] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Received event network-changed-e21d7ddc-a0d3-4026-a937-f392d5e20fa4 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.745688] env[61243]: DEBUG nova.compute.manager [req-33f05a7b-f106-43c5-83ca-ff3e237d392c req-530fc4cf-32e3-4189-8ee7-048faa0db130 service nova] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Refreshing instance network info cache due to event network-changed-e21d7ddc-a0d3-4026-a937-f392d5e20fa4. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 857.746046] env[61243]: DEBUG oslo_concurrency.lockutils [req-33f05a7b-f106-43c5-83ca-ff3e237d392c req-530fc4cf-32e3-4189-8ee7-048faa0db130 service nova] Acquiring lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.746356] env[61243]: DEBUG oslo_concurrency.lockutils [req-33f05a7b-f106-43c5-83ca-ff3e237d392c req-530fc4cf-32e3-4189-8ee7-048faa0db130 service nova] Acquired lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.746724] env[61243]: DEBUG nova.network.neutron [req-33f05a7b-f106-43c5-83ca-ff3e237d392c req-530fc4cf-32e3-4189-8ee7-048faa0db130 service nova] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Refreshing network info cache for port e21d7ddc-a0d3-4026-a937-f392d5e20fa4 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 857.814802] env[61243]: DEBUG oslo_concurrency.lockutils [None req-73e4d04e-82db-4707-a62f-c0ded2f7327b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.950s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.878789] env[61243]: DEBUG oslo_vmware.api [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1338940, 'name': PowerOnVM_Task, 'duration_secs': 0.666581} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.881890] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.882286] env[61243]: INFO nova.compute.manager [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Took 7.74 seconds to spawn the instance on the hypervisor. [ 857.882579] env[61243]: DEBUG nova.compute.manager [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.884356] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47cd3169-80a5-4c5f-bd3d-b06da36068b6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.910633] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "refresh_cache-5daccb48-b9bf-423a-ab1a-5aa26d8bec87" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.910890] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquired lock "refresh_cache-5daccb48-b9bf-423a-ab1a-5aa26d8bec87" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.911298] env[61243]: DEBUG nova.network.neutron [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 857.915816] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d108af-5c11-44a4-96b0-251421b7059b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.925567] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6587ab-5ccc-488d-86ca-7836a2eba8aa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.968276] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de32260a-2618-4b95-97a4-7b0005e6c659 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.978019] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': task-1338941, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519711} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.980581] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] f73a4d39-5478-4135-9be6-e59f3e29788d/f73a4d39-5478-4135-9be6-e59f3e29788d.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 857.980581] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 857.981019] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-027d6d05-f2a8-4c65-9750-b1fd5c50d269 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.983483] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b9cf0ae-6679-4911-9b91-33b95eaaf715 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.999675] env[61243]: DEBUG nova.compute.provider_tree [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 858.004259] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Waiting for the task: (returnval){ [ 858.004259] env[61243]: value = "task-1338943" [ 858.004259] env[61243]: _type = "Task" [ 858.004259] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.013304] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': task-1338943, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.226387] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 858.226387] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52df14fb-4eb5-380a-7937-fd0eb03b8a7b" [ 858.226387] env[61243]: _type = "HttpNfcLease" [ 858.226387] env[61243]: } is ready. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 858.227053] env[61243]: DEBUG oslo_vmware.rw_handles [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 858.227053] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52df14fb-4eb5-380a-7937-fd0eb03b8a7b" [ 858.227053] env[61243]: _type = "HttpNfcLease" [ 858.227053] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 858.227514] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b05a24-b19f-47be-a05e-4fc1ca17452e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.236301] env[61243]: DEBUG oslo_vmware.rw_handles [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527d22a4-7427-d947-4e73-f46c6798cdea/disk-0.vmdk from lease info. {{(pid=61243) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 858.236490] env[61243]: DEBUG oslo_vmware.rw_handles [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527d22a4-7427-d947-4e73-f46c6798cdea/disk-0.vmdk for reading. {{(pid=61243) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 858.328422] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-49a2439d-290d-4c06-bc68-9213f8647e98 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.404594] env[61243]: INFO nova.compute.manager [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Took 25.99 seconds to build instance. [ 858.455529] env[61243]: DEBUG nova.network.neutron [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 858.519025] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': task-1338943, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056696} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.519025] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 858.519965] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6104f76c-131e-4b6b-840d-ff14ffba0fbe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.541531] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] f73a4d39-5478-4135-9be6-e59f3e29788d/f73a4d39-5478-4135-9be6-e59f3e29788d.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 858.543767] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca3eba88-8d58-44c6-8d0e-52c2b89c1423 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.557851] env[61243]: DEBUG nova.scheduler.client.report [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 91 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 858.558103] env[61243]: DEBUG nova.compute.provider_tree [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 91 to 92 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 858.558288] env[61243]: DEBUG nova.compute.provider_tree [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 858.572141] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Waiting for the task: (returnval){ [ 858.572141] env[61243]: value = "task-1338944" [ 858.572141] env[61243]: _type = "Task" [ 858.572141] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.583289] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': task-1338944, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.627554] env[61243]: DEBUG nova.network.neutron [req-33f05a7b-f106-43c5-83ca-ff3e237d392c req-530fc4cf-32e3-4189-8ee7-048faa0db130 service nova] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updated VIF entry in instance network info cache for port e21d7ddc-a0d3-4026-a937-f392d5e20fa4. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 858.628111] env[61243]: DEBUG nova.network.neutron [req-33f05a7b-f106-43c5-83ca-ff3e237d392c req-530fc4cf-32e3-4189-8ee7-048faa0db130 service nova] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating instance_info_cache with network_info: [{"id": "e21d7ddc-a0d3-4026-a937-f392d5e20fa4", "address": "fa:16:3e:56:77:58", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape21d7ddc-a0", "ovs_interfaceid": "e21d7ddc-a0d3-4026-a937-f392d5e20fa4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.667643] env[61243]: DEBUG nova.network.neutron [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Updating instance_info_cache with network_info: [{"id": "1aa2d7ff-eac6-472b-b7a7-57c542d3f111", "address": "fa:16:3e:38:41:0f", "network": {"id": "49be247d-97e7-4085-8272-4ccf41767d10", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1335943921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "399030c8863346ea97f6da669ff3868b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aa2d7ff-ea", "ovs_interfaceid": "1aa2d7ff-eac6-472b-b7a7-57c542d3f111", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.906755] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8d2d4585-9853-4615-b650-bf8c58ff89d9 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "c1a0d242-4eae-4c03-8341-840b41341f17" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.013s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.063605] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.064653] env[61243]: DEBUG nova.compute.manager [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 859.068025] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.412s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.068289] env[61243]: DEBUG nova.objects.instance [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lazy-loading 'resources' on Instance uuid 36e55334-8628-4dd7-a845-f4ae3d8e7ff9 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 859.083154] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': task-1338944, 'name': ReconfigVM_Task, 'duration_secs': 0.496797} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.084386] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Reconfigured VM instance instance-00000047 to attach disk [datastore1] f73a4d39-5478-4135-9be6-e59f3e29788d/f73a4d39-5478-4135-9be6-e59f3e29788d.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 859.085126] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39b022f4-6070-4b8d-aafd-5bd1bf938839 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.091283] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Waiting for the task: (returnval){ [ 859.091283] env[61243]: value = "task-1338945" [ 859.091283] env[61243]: _type = "Task" [ 859.091283] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.100177] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': task-1338945, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.134041] env[61243]: DEBUG oslo_concurrency.lockutils [req-33f05a7b-f106-43c5-83ca-ff3e237d392c req-530fc4cf-32e3-4189-8ee7-048faa0db130 service nova] Releasing lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.145145] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "07ee984d-476e-484d-ba80-0ec2e411faa9" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.145464] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.170887] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Releasing lock "refresh_cache-5daccb48-b9bf-423a-ab1a-5aa26d8bec87" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.171785] env[61243]: DEBUG nova.compute.manager [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Instance network_info: |[{"id": "1aa2d7ff-eac6-472b-b7a7-57c542d3f111", "address": "fa:16:3e:38:41:0f", "network": {"id": "49be247d-97e7-4085-8272-4ccf41767d10", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1335943921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "399030c8863346ea97f6da669ff3868b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aa2d7ff-ea", "ovs_interfaceid": "1aa2d7ff-eac6-472b-b7a7-57c542d3f111", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 859.171924] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:41:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e547d234-640c-449b-8279-0b16f75d6627', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1aa2d7ff-eac6-472b-b7a7-57c542d3f111', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 859.180500] env[61243]: DEBUG oslo.service.loopingcall [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.181235] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 859.181502] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d4b006e2-ead0-4b35-aabf-dbe510e18e46 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.198190] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.199099] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.205591] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.205591] env[61243]: value = "task-1338946" [ 859.205591] env[61243]: _type = "Task" [ 859.205591] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.215864] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338946, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.504488] env[61243]: DEBUG nova.compute.manager [req-18a596f0-f5c2-40c8-b305-c83994d4a6dc req-7d9f4fd9-84b2-4f21-a346-4dbf10eac614 service nova] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Received event network-changed-1aa2d7ff-eac6-472b-b7a7-57c542d3f111 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.504488] env[61243]: DEBUG nova.compute.manager [req-18a596f0-f5c2-40c8-b305-c83994d4a6dc req-7d9f4fd9-84b2-4f21-a346-4dbf10eac614 service nova] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Refreshing instance network info cache due to event network-changed-1aa2d7ff-eac6-472b-b7a7-57c542d3f111. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 859.504488] env[61243]: DEBUG oslo_concurrency.lockutils [req-18a596f0-f5c2-40c8-b305-c83994d4a6dc req-7d9f4fd9-84b2-4f21-a346-4dbf10eac614 service nova] Acquiring lock "refresh_cache-5daccb48-b9bf-423a-ab1a-5aa26d8bec87" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.504697] env[61243]: DEBUG oslo_concurrency.lockutils [req-18a596f0-f5c2-40c8-b305-c83994d4a6dc req-7d9f4fd9-84b2-4f21-a346-4dbf10eac614 service nova] Acquired lock "refresh_cache-5daccb48-b9bf-423a-ab1a-5aa26d8bec87" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.504739] env[61243]: DEBUG nova.network.neutron [req-18a596f0-f5c2-40c8-b305-c83994d4a6dc req-7d9f4fd9-84b2-4f21-a346-4dbf10eac614 service nova] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Refreshing network info cache for port 1aa2d7ff-eac6-472b-b7a7-57c542d3f111 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 859.569851] env[61243]: DEBUG nova.compute.utils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 859.572119] env[61243]: DEBUG nova.compute.manager [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 859.572555] env[61243]: DEBUG nova.network.neutron [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 859.605476] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': task-1338945, 'name': Rename_Task, 'duration_secs': 0.261765} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.608746] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 859.609347] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-031547d7-c6a5-41d2-a280-d42d23db44f8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.616227] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Waiting for the task: (returnval){ [ 859.616227] env[61243]: value = "task-1338947" [ 859.616227] env[61243]: _type = "Task" [ 859.616227] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.623940] env[61243]: DEBUG nova.policy [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06e247d66c8d43218302a35ead439405', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e641e72e4d2048218ef403456dbb2b82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 859.631909] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': task-1338947, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.649211] env[61243]: DEBUG nova.compute.utils [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 859.713835] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.713835] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Starting heal instance info cache {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 859.714141] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Rebuilding the list of instances to heal {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 859.721227] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338946, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.897917] env[61243]: DEBUG nova.compute.manager [req-68ab4f2a-7b09-4194-8198-850281d7512a req-31b892da-d883-46f0-a42a-47b3abbb49eb service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Received event network-changed-970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.898293] env[61243]: DEBUG nova.compute.manager [req-68ab4f2a-7b09-4194-8198-850281d7512a req-31b892da-d883-46f0-a42a-47b3abbb49eb service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Refreshing instance network info cache due to event network-changed-970e8509-164d-4ddf-8f0c-795e92883f6b. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 859.898517] env[61243]: DEBUG oslo_concurrency.lockutils [req-68ab4f2a-7b09-4194-8198-850281d7512a req-31b892da-d883-46f0-a42a-47b3abbb49eb service nova] Acquiring lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.898677] env[61243]: DEBUG oslo_concurrency.lockutils [req-68ab4f2a-7b09-4194-8198-850281d7512a req-31b892da-d883-46f0-a42a-47b3abbb49eb service nova] Acquired lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.898859] env[61243]: DEBUG nova.network.neutron [req-68ab4f2a-7b09-4194-8198-850281d7512a req-31b892da-d883-46f0-a42a-47b3abbb49eb service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Refreshing network info cache for port 970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 859.940575] env[61243]: DEBUG nova.network.neutron [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Successfully created port: fe33c43c-f24f-4b9e-9cb9-577bc487106e {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 859.995170] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e9c8722-bc79-4c7a-af77-4085607d0bbf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.007503] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec34e767-e39b-476e-ae4f-55cd96a510f1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.049150] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95130f0-1dfb-42bb-bfd2-565582912fbd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.061996] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f37583-f591-4c6e-9763-875d2d6c31af {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.079480] env[61243]: DEBUG nova.compute.provider_tree [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.081386] env[61243]: DEBUG nova.compute.manager [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 860.128933] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': task-1338947, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.151866] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.222431] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Skipping network cache update for instance because it is Building. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 860.222742] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Skipping network cache update for instance because it is Building. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 860.223134] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Skipping network cache update for instance because it is Building. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 860.233220] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338946, 'name': CreateVM_Task, 'duration_secs': 0.546279} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.234100] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 860.235294] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.235596] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.236047] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 860.236457] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd61b230-5c57-420d-b8aa-5d143dd613eb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.242353] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 860.242353] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d7e8ac-9230-386c-1f6c-7803bd8a3a53" [ 860.242353] env[61243]: _type = "Task" [ 860.242353] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.252425] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d7e8ac-9230-386c-1f6c-7803bd8a3a53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.510458] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.510693] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquired lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.510936] env[61243]: DEBUG nova.network.neutron [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Forcefully refreshing network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 860.511195] env[61243]: DEBUG nova.objects.instance [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lazy-loading 'info_cache' on Instance uuid 86179fb0-99df-4b10-a815-c19168e9521e {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.585856] env[61243]: DEBUG nova.scheduler.client.report [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.630552] env[61243]: DEBUG oslo_vmware.api [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': task-1338947, 'name': PowerOnVM_Task, 'duration_secs': 0.774857} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.630928] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 860.631203] env[61243]: INFO nova.compute.manager [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Took 8.07 seconds to spawn the instance on the hypervisor. [ 860.631894] env[61243]: DEBUG nova.compute.manager [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 860.633374] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7760f7c-f27b-453e-b9b8-5f2da1c91e91 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.686415] env[61243]: DEBUG nova.network.neutron [req-18a596f0-f5c2-40c8-b305-c83994d4a6dc req-7d9f4fd9-84b2-4f21-a346-4dbf10eac614 service nova] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Updated VIF entry in instance network info cache for port 1aa2d7ff-eac6-472b-b7a7-57c542d3f111. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 860.686939] env[61243]: DEBUG nova.network.neutron [req-18a596f0-f5c2-40c8-b305-c83994d4a6dc req-7d9f4fd9-84b2-4f21-a346-4dbf10eac614 service nova] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Updating instance_info_cache with network_info: [{"id": "1aa2d7ff-eac6-472b-b7a7-57c542d3f111", "address": "fa:16:3e:38:41:0f", "network": {"id": "49be247d-97e7-4085-8272-4ccf41767d10", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1335943921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "399030c8863346ea97f6da669ff3868b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aa2d7ff-ea", "ovs_interfaceid": "1aa2d7ff-eac6-472b-b7a7-57c542d3f111", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.758575] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d7e8ac-9230-386c-1f6c-7803bd8a3a53, 'name': SearchDatastore_Task, 'duration_secs': 0.01279} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.759312] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.759679] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.759901] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.760074] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.760965] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.760965] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9fdd8712-eb5a-4642-bfa1-2dfc033da06a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.770453] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.770651] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 860.771690] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b85387b3-602f-499e-9d3d-b6cec63c76cf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.778275] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 860.778275] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521395bb-8b57-dfd6-74c8-08f523fc8d61" [ 860.778275] env[61243]: _type = "Task" [ 860.778275] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.789478] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521395bb-8b57-dfd6-74c8-08f523fc8d61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.816790] env[61243]: DEBUG nova.network.neutron [req-68ab4f2a-7b09-4194-8198-850281d7512a req-31b892da-d883-46f0-a42a-47b3abbb49eb service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Updated VIF entry in instance network info cache for port 970e8509-164d-4ddf-8f0c-795e92883f6b. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 860.817200] env[61243]: DEBUG nova.network.neutron [req-68ab4f2a-7b09-4194-8198-850281d7512a req-31b892da-d883-46f0-a42a-47b3abbb49eb service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Updating instance_info_cache with network_info: [{"id": "970e8509-164d-4ddf-8f0c-795e92883f6b", "address": "fa:16:3e:00:af:7d", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap970e8509-16", "ovs_interfaceid": "970e8509-164d-4ddf-8f0c-795e92883f6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.096824] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.029s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.099895] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.676s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.100488] env[61243]: DEBUG nova.objects.instance [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Lazy-loading 'resources' on Instance uuid 6cee3dd1-c13d-427d-889c-6cd3ae388a31 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 861.104037] env[61243]: DEBUG nova.compute.manager [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 861.124662] env[61243]: INFO nova.scheduler.client.report [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Deleted allocations for instance 36e55334-8628-4dd7-a845-f4ae3d8e7ff9 [ 861.128572] env[61243]: DEBUG nova.virt.hardware [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 861.129117] env[61243]: DEBUG nova.virt.hardware [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 861.129117] env[61243]: DEBUG nova.virt.hardware [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 861.129185] env[61243]: DEBUG nova.virt.hardware [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 861.129295] env[61243]: DEBUG nova.virt.hardware [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 861.129447] env[61243]: DEBUG nova.virt.hardware [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 861.129664] env[61243]: DEBUG nova.virt.hardware [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 861.129914] env[61243]: DEBUG nova.virt.hardware [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 861.130142] env[61243]: DEBUG nova.virt.hardware [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 861.130282] env[61243]: DEBUG nova.virt.hardware [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 861.130458] env[61243]: DEBUG nova.virt.hardware [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.131707] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899af61a-1955-44e6-8146-b7765bff8763 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.144158] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f6343a-ebe0-4aed-81d7-53961ed25883 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.166807] env[61243]: INFO nova.compute.manager [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Took 22.88 seconds to build instance. [ 861.190778] env[61243]: DEBUG oslo_concurrency.lockutils [req-18a596f0-f5c2-40c8-b305-c83994d4a6dc req-7d9f4fd9-84b2-4f21-a346-4dbf10eac614 service nova] Releasing lock "refresh_cache-5daccb48-b9bf-423a-ab1a-5aa26d8bec87" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.252151] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "07ee984d-476e-484d-ba80-0ec2e411faa9" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.252426] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.252671] env[61243]: INFO nova.compute.manager [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Attaching volume 3111ce1e-7b01-4717-885c-55baf54556d1 to /dev/sdb [ 861.293136] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521395bb-8b57-dfd6-74c8-08f523fc8d61, 'name': SearchDatastore_Task, 'duration_secs': 0.010854} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.295242] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-651d4549-e134-4d6c-8905-22f31e505474 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.297789] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c849c05-5669-496b-a6a7-c34f3d0a0111 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.305446] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5150263b-228a-48b6-a457-7506a784bf88 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.308324] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 861.308324] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52917884-574d-4383-4ecb-8800491e9bb7" [ 861.308324] env[61243]: _type = "Task" [ 861.308324] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.317743] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52917884-574d-4383-4ecb-8800491e9bb7, 'name': SearchDatastore_Task, 'duration_secs': 0.010563} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.322368] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.322665] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 5daccb48-b9bf-423a-ab1a-5aa26d8bec87/5daccb48-b9bf-423a-ab1a-5aa26d8bec87.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 861.323670] env[61243]: DEBUG oslo_concurrency.lockutils [req-68ab4f2a-7b09-4194-8198-850281d7512a req-31b892da-d883-46f0-a42a-47b3abbb49eb service nova] Releasing lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.324875] env[61243]: DEBUG nova.virt.block_device [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Updating existing volume attachment record: d01337df-ef69-436c-ae83-11476c7a69be {{(pid=61243) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 861.328234] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f7de0c58-b51a-465f-998c-db232f47e1b1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.336464] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 861.336464] env[61243]: value = "task-1338948" [ 861.336464] env[61243]: _type = "Task" [ 861.336464] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.348043] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338948, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.645995] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ab7a0ded-9b6d-4960-a713-96eddf62e813 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "36e55334-8628-4dd7-a845-f4ae3d8e7ff9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.628s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.670786] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6638576a-ddc9-4892-8bc9-e1dcdee9e7b9 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Lock "f73a4d39-5478-4135-9be6-e59f3e29788d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.553s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.848676] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338948, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.972453] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf06b9e8-b46e-4049-998a-dd4a0b04a9b6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.980495] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f9c90c8-6b93-4cb6-a968-09f38c3106dd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.022254] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de685834-756b-44d3-a470-0a3f7b731743 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.029657] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbdaa5a1-0103-4c5d-a2a4-d63fd8960769 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.045215] env[61243]: DEBUG nova.compute.provider_tree [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.075751] env[61243]: DEBUG nova.compute.manager [req-99ae8fb5-bff4-4f38-ac7d-290001e7ef00 req-b3c71a8e-8cc4-4dee-b857-ec4322188de8 service nova] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Received event network-vif-plugged-fe33c43c-f24f-4b9e-9cb9-577bc487106e {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.076064] env[61243]: DEBUG oslo_concurrency.lockutils [req-99ae8fb5-bff4-4f38-ac7d-290001e7ef00 req-b3c71a8e-8cc4-4dee-b857-ec4322188de8 service nova] Acquiring lock "ca230425-aaa7-4c32-81d3-379cbce0d41b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.076258] env[61243]: DEBUG oslo_concurrency.lockutils [req-99ae8fb5-bff4-4f38-ac7d-290001e7ef00 req-b3c71a8e-8cc4-4dee-b857-ec4322188de8 service nova] Lock "ca230425-aaa7-4c32-81d3-379cbce0d41b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.076456] env[61243]: DEBUG oslo_concurrency.lockutils [req-99ae8fb5-bff4-4f38-ac7d-290001e7ef00 req-b3c71a8e-8cc4-4dee-b857-ec4322188de8 service nova] Lock "ca230425-aaa7-4c32-81d3-379cbce0d41b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.076671] env[61243]: DEBUG nova.compute.manager [req-99ae8fb5-bff4-4f38-ac7d-290001e7ef00 req-b3c71a8e-8cc4-4dee-b857-ec4322188de8 service nova] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] No waiting events found dispatching network-vif-plugged-fe33c43c-f24f-4b9e-9cb9-577bc487106e {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 862.076964] env[61243]: WARNING nova.compute.manager [req-99ae8fb5-bff4-4f38-ac7d-290001e7ef00 req-b3c71a8e-8cc4-4dee-b857-ec4322188de8 service nova] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Received unexpected event network-vif-plugged-fe33c43c-f24f-4b9e-9cb9-577bc487106e for instance with vm_state building and task_state spawning. [ 862.230725] env[61243]: DEBUG nova.network.neutron [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Updating instance_info_cache with network_info: [{"id": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "address": "fa:16:3e:b0:f1:66", "network": {"id": "49be247d-97e7-4085-8272-4ccf41767d10", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1335943921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "399030c8863346ea97f6da669ff3868b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2248ec9-02", "ovs_interfaceid": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.348656] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338948, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.674501} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.348995] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 5daccb48-b9bf-423a-ab1a-5aa26d8bec87/5daccb48-b9bf-423a-ab1a-5aa26d8bec87.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 862.349298] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.349607] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4d7e6fb6-b024-4261-8ca8-ba5e887f59ab {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.356303] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 862.356303] env[61243]: value = "task-1338952" [ 862.356303] env[61243]: _type = "Task" [ 862.356303] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.364380] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338952, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.430523] env[61243]: DEBUG oslo_concurrency.lockutils [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "interface-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.430919] env[61243]: DEBUG oslo_concurrency.lockutils [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.431378] env[61243]: DEBUG nova.objects.instance [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lazy-loading 'flavor' on Instance uuid 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.466876] env[61243]: DEBUG nova.network.neutron [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Successfully updated port: fe33c43c-f24f-4b9e-9cb9-577bc487106e {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 862.549038] env[61243]: DEBUG nova.scheduler.client.report [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 862.733757] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Releasing lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.734023] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Updated the network info_cache for instance {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 862.734275] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 862.734464] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 862.734647] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 862.734834] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 862.735015] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 862.735208] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 862.735417] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61243) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 862.735507] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager.update_available_resource {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 862.752978] env[61243]: DEBUG nova.compute.manager [req-ad8160fd-063d-45cf-9298-f26b2ed7d205 req-45e7e520-f5c5-4b77-b403-2dfe1bfe9231 service nova] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Received event network-changed-fe33c43c-f24f-4b9e-9cb9-577bc487106e {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.752978] env[61243]: DEBUG nova.compute.manager [req-ad8160fd-063d-45cf-9298-f26b2ed7d205 req-45e7e520-f5c5-4b77-b403-2dfe1bfe9231 service nova] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Refreshing instance network info cache due to event network-changed-fe33c43c-f24f-4b9e-9cb9-577bc487106e. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 862.753208] env[61243]: DEBUG oslo_concurrency.lockutils [req-ad8160fd-063d-45cf-9298-f26b2ed7d205 req-45e7e520-f5c5-4b77-b403-2dfe1bfe9231 service nova] Acquiring lock "refresh_cache-ca230425-aaa7-4c32-81d3-379cbce0d41b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.753355] env[61243]: DEBUG oslo_concurrency.lockutils [req-ad8160fd-063d-45cf-9298-f26b2ed7d205 req-45e7e520-f5c5-4b77-b403-2dfe1bfe9231 service nova] Acquired lock "refresh_cache-ca230425-aaa7-4c32-81d3-379cbce0d41b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.753533] env[61243]: DEBUG nova.network.neutron [req-ad8160fd-063d-45cf-9298-f26b2ed7d205 req-45e7e520-f5c5-4b77-b403-2dfe1bfe9231 service nova] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Refreshing network info cache for port fe33c43c-f24f-4b9e-9cb9-577bc487106e {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 862.867625] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338952, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076857} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.867964] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.868813] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d21cf7f9-fc94-4c73-92ef-6e11a2aa14af {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.894064] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] 5daccb48-b9bf-423a-ab1a-5aa26d8bec87/5daccb48-b9bf-423a-ab1a-5aa26d8bec87.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.894850] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5970e399-d562-47bc-bdff-813758df4180 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.915659] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 862.915659] env[61243]: value = "task-1338953" [ 862.915659] env[61243]: _type = "Task" [ 862.915659] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.924718] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338953, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.969965] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "refresh_cache-ca230425-aaa7-4c32-81d3-379cbce0d41b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.012617] env[61243]: DEBUG nova.objects.instance [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lazy-loading 'pci_requests' on Instance uuid 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.054031] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.954s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.057322] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.606s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.059779] env[61243]: INFO nova.compute.claims [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 863.081823] env[61243]: INFO nova.scheduler.client.report [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Deleted allocations for instance 6cee3dd1-c13d-427d-889c-6cd3ae388a31 [ 863.239331] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.288453] env[61243]: DEBUG nova.network.neutron [req-ad8160fd-063d-45cf-9298-f26b2ed7d205 req-45e7e520-f5c5-4b77-b403-2dfe1bfe9231 service nova] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 863.381227] env[61243]: DEBUG nova.network.neutron [req-ad8160fd-063d-45cf-9298-f26b2ed7d205 req-45e7e520-f5c5-4b77-b403-2dfe1bfe9231 service nova] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.425342] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338953, 'name': ReconfigVM_Task, 'duration_secs': 0.347377} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.425638] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Reconfigured VM instance instance-00000048 to attach disk [datastore1] 5daccb48-b9bf-423a-ab1a-5aa26d8bec87/5daccb48-b9bf-423a-ab1a-5aa26d8bec87.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 863.426317] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9566ed13-0878-4713-8446-22675d347665 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.432857] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 863.432857] env[61243]: value = "task-1338954" [ 863.432857] env[61243]: _type = "Task" [ 863.432857] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.441185] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338954, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.515917] env[61243]: DEBUG nova.objects.base [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Object Instance<7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00> lazy-loaded attributes: flavor,pci_requests {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 863.515917] env[61243]: DEBUG nova.network.neutron [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 863.557197] env[61243]: DEBUG nova.policy [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b32f4788825f4a8e82e028cc10347746', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '683c027e4adc4cc4a6b8747a8c8c0de9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 863.589445] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4222e73c-3da1-447d-b26c-53011abbb49e tempest-ServerShowV254Test-1003586897 tempest-ServerShowV254Test-1003586897-project-member] Lock "6cee3dd1-c13d-427d-889c-6cd3ae388a31" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.921s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.634335] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Acquiring lock "f73a4d39-5478-4135-9be6-e59f3e29788d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.634607] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Lock "f73a4d39-5478-4135-9be6-e59f3e29788d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.634825] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Acquiring lock "f73a4d39-5478-4135-9be6-e59f3e29788d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.635033] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Lock "f73a4d39-5478-4135-9be6-e59f3e29788d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.635338] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Lock "f73a4d39-5478-4135-9be6-e59f3e29788d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.637638] env[61243]: INFO nova.compute.manager [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Terminating instance [ 863.639636] env[61243]: DEBUG nova.compute.manager [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 863.639840] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 863.640703] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9afb147b-676f-4105-b51d-e97747200f2d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.652899] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 863.652899] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fbc46042-16aa-4791-8443-f8f7e02c3e21 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.655141] env[61243]: DEBUG oslo_vmware.api [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Waiting for the task: (returnval){ [ 863.655141] env[61243]: value = "task-1338955" [ 863.655141] env[61243]: _type = "Task" [ 863.655141] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.663369] env[61243]: DEBUG oslo_vmware.api [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': task-1338955, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.876750] env[61243]: DEBUG nova.network.neutron [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Successfully created port: 7138c7dd-4984-4d4f-a631-0d0f531c0044 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 863.883763] env[61243]: DEBUG oslo_concurrency.lockutils [req-ad8160fd-063d-45cf-9298-f26b2ed7d205 req-45e7e520-f5c5-4b77-b403-2dfe1bfe9231 service nova] Releasing lock "refresh_cache-ca230425-aaa7-4c32-81d3-379cbce0d41b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.884215] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "refresh_cache-ca230425-aaa7-4c32-81d3-379cbce0d41b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.884382] env[61243]: DEBUG nova.network.neutron [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 863.948023] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338954, 'name': Rename_Task, 'duration_secs': 0.257167} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.948023] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 863.948023] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b7615a2-4752-441e-8742-9a823873c25e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.956519] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 863.956519] env[61243]: value = "task-1338957" [ 863.956519] env[61243]: _type = "Task" [ 863.956519] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.967646] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338957, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.171317] env[61243]: DEBUG oslo_vmware.api [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': task-1338955, 'name': PowerOffVM_Task, 'duration_secs': 0.260635} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.175074] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 864.175403] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 864.175960] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-42fc084b-e69f-493f-8b70-78e8e18dffc8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.248835] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 864.248987] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 864.249292] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Deleting the datastore file [datastore1] f73a4d39-5478-4135-9be6-e59f3e29788d {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.249636] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a0a3730-89e9-4a2c-8d8c-7971327ba23a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.258014] env[61243]: DEBUG oslo_vmware.api [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Waiting for the task: (returnval){ [ 864.258014] env[61243]: value = "task-1338959" [ 864.258014] env[61243]: _type = "Task" [ 864.258014] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.266931] env[61243]: DEBUG oslo_vmware.api [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': task-1338959, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.386564] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-957ef91a-6426-45e0-8b45-35ca9315d75d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.400255] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9890be55-2b1b-43b3-bd66-366f5fd5f82e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.433203] env[61243]: DEBUG nova.network.neutron [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 864.437034] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-938055b9-daa3-4bf1-983f-d9bf26ad01b9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.444055] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f260af5-3e62-4db0-8a87-479d76bb8103 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.458378] env[61243]: DEBUG nova.compute.provider_tree [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.469963] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338957, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.611810] env[61243]: DEBUG nova.network.neutron [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Updating instance_info_cache with network_info: [{"id": "fe33c43c-f24f-4b9e-9cb9-577bc487106e", "address": "fa:16:3e:43:4a:58", "network": {"id": "bce1aecb-e043-4f68-99e2-c6b03c4bb866", "bridge": "br-int", "label": "tempest-ServersTestJSON-652904947-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e641e72e4d2048218ef403456dbb2b82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe33c43c-f2", "ovs_interfaceid": "fe33c43c-f24f-4b9e-9cb9-577bc487106e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.767668] env[61243]: DEBUG oslo_vmware.api [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Task: {'id': task-1338959, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.231376} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.768013] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.768177] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 864.768361] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 864.768541] env[61243]: INFO nova.compute.manager [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 864.768786] env[61243]: DEBUG oslo.service.loopingcall [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.768985] env[61243]: DEBUG nova.compute.manager [-] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 864.769095] env[61243]: DEBUG nova.network.neutron [-] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 864.965959] env[61243]: DEBUG nova.scheduler.client.report [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.972759] env[61243]: DEBUG oslo_vmware.api [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338957, 'name': PowerOnVM_Task, 'duration_secs': 0.568844} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.973279] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 864.973449] env[61243]: INFO nova.compute.manager [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Took 7.87 seconds to spawn the instance on the hypervisor. [ 864.973629] env[61243]: DEBUG nova.compute.manager [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 864.974396] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03734b1-7806-4491-a7ea-6db226325b51 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.103397] env[61243]: DEBUG nova.compute.manager [req-e9354632-6bc6-49db-b168-f5ecc01cb3a9 req-9b420400-da70-48af-90f4-6962fdb98454 service nova] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Received event network-vif-deleted-c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.103397] env[61243]: INFO nova.compute.manager [req-e9354632-6bc6-49db-b168-f5ecc01cb3a9 req-9b420400-da70-48af-90f4-6962fdb98454 service nova] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Neutron deleted interface c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3; detaching it from the instance and deleting it from the info cache [ 865.103546] env[61243]: DEBUG nova.network.neutron [req-e9354632-6bc6-49db-b168-f5ecc01cb3a9 req-9b420400-da70-48af-90f4-6962fdb98454 service nova] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.113746] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "refresh_cache-ca230425-aaa7-4c32-81d3-379cbce0d41b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.114088] env[61243]: DEBUG nova.compute.manager [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Instance network_info: |[{"id": "fe33c43c-f24f-4b9e-9cb9-577bc487106e", "address": "fa:16:3e:43:4a:58", "network": {"id": "bce1aecb-e043-4f68-99e2-c6b03c4bb866", "bridge": "br-int", "label": "tempest-ServersTestJSON-652904947-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e641e72e4d2048218ef403456dbb2b82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe33c43c-f2", "ovs_interfaceid": "fe33c43c-f24f-4b9e-9cb9-577bc487106e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 865.114485] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:4a:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '90878b7b-ddb7-4f47-892b-d6e06f73475f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fe33c43c-f24f-4b9e-9cb9-577bc487106e', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 865.121925] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Creating folder: Project (e641e72e4d2048218ef403456dbb2b82). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 865.122605] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-15067cc6-4920-434f-ab7c-079f0a3358d6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.132916] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Created folder: Project (e641e72e4d2048218ef403456dbb2b82) in parent group-v285636. [ 865.133119] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Creating folder: Instances. Parent ref: group-v285712. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 865.134038] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0095f953-937f-4a72-90c6-b02ff72dc511 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.141252] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Created folder: Instances in parent group-v285712. [ 865.141487] env[61243]: DEBUG oslo.service.loopingcall [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.141678] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 865.141879] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ab817e73-210a-489a-8714-8f434cd23f26 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.161322] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 865.161322] env[61243]: value = "task-1338962" [ 865.161322] env[61243]: _type = "Task" [ 865.161322] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.168745] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338962, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.429182] env[61243]: DEBUG nova.network.neutron [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Successfully updated port: 7138c7dd-4984-4d4f-a631-0d0f531c0044 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 865.474676] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.475235] env[61243]: DEBUG nova.compute.manager [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 865.478267] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.045s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.479795] env[61243]: INFO nova.compute.claims [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 865.490791] env[61243]: INFO nova.compute.manager [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Took 23.29 seconds to build instance. [ 865.503083] env[61243]: DEBUG nova.network.neutron [-] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.607987] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e8741203-f746-4d32-977d-f15c96c772df {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.617376] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b941e5-1e49-42a7-8083-d1ed924ced4a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.648610] env[61243]: DEBUG nova.compute.manager [req-e9354632-6bc6-49db-b168-f5ecc01cb3a9 req-9b420400-da70-48af-90f4-6962fdb98454 service nova] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Detach interface failed, port_id=c640e6df-911c-4d0b-a5cc-b70ca4eb9ae3, reason: Instance f73a4d39-5478-4135-9be6-e59f3e29788d could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 865.671162] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338962, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.889351] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Volume attach. Driver type: vmdk {{(pid=61243) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 865.889611] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285711', 'volume_id': '3111ce1e-7b01-4717-885c-55baf54556d1', 'name': 'volume-3111ce1e-7b01-4717-885c-55baf54556d1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '07ee984d-476e-484d-ba80-0ec2e411faa9', 'attached_at': '', 'detached_at': '', 'volume_id': '3111ce1e-7b01-4717-885c-55baf54556d1', 'serial': '3111ce1e-7b01-4717-885c-55baf54556d1'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 865.890494] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6499e6-8373-44af-9c8e-ea834e9efa54 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.909255] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fde92c4-90db-48ef-a848-b609915fbb54 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.934396] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] volume-3111ce1e-7b01-4717-885c-55baf54556d1/volume-3111ce1e-7b01-4717-885c-55baf54556d1.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.935138] env[61243]: DEBUG oslo_concurrency.lockutils [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.935314] env[61243]: DEBUG oslo_concurrency.lockutils [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.935734] env[61243]: DEBUG nova.network.neutron [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 865.936597] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a88e2abf-e149-4253-8812-ce51c84af40c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.957699] env[61243]: DEBUG oslo_vmware.api [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 865.957699] env[61243]: value = "task-1338963" [ 865.957699] env[61243]: _type = "Task" [ 865.957699] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.966708] env[61243]: DEBUG oslo_vmware.api [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338963, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.985453] env[61243]: DEBUG nova.compute.utils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 865.990429] env[61243]: WARNING nova.network.neutron [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] 7724b87f-3721-42a5-81ce-27ae98fc99c2 already exists in list: networks containing: ['7724b87f-3721-42a5-81ce-27ae98fc99c2']. ignoring it [ 865.992743] env[61243]: DEBUG nova.compute.manager [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 865.993053] env[61243]: DEBUG nova.network.neutron [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 865.995950] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c6f2a-4ea3-464c-87ef-caa0161c1e3f tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "5daccb48-b9bf-423a-ab1a-5aa26d8bec87" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.513s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.005941] env[61243]: INFO nova.compute.manager [-] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Took 1.24 seconds to deallocate network for instance. [ 866.072672] env[61243]: DEBUG nova.policy [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc450f796a9a49cda4328efe45c92057', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bf51131cc0e247669ddf5495ffebd8d5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 866.175042] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338962, 'name': CreateVM_Task, 'duration_secs': 0.580844} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.175042] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 866.175042] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.175042] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.175042] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 866.175042] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa92454c-426f-4303-9f83-ec5919ba6ce8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.179023] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 866.179023] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f44415-7602-738a-7da1-33a48887bd61" [ 866.179023] env[61243]: _type = "Task" [ 866.179023] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.186876] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f44415-7602-738a-7da1-33a48887bd61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.472007] env[61243]: DEBUG oslo_vmware.api [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338963, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.490864] env[61243]: DEBUG nova.compute.manager [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 866.514289] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.516338] env[61243]: DEBUG nova.network.neutron [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Updating instance_info_cache with network_info: [{"id": "4cd8c409-8daa-4f53-ac5f-d2ac25de247b", "address": "fa:16:3e:11:60:2d", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4cd8c409-8d", "ovs_interfaceid": "4cd8c409-8daa-4f53-ac5f-d2ac25de247b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7138c7dd-4984-4d4f-a631-0d0f531c0044", "address": "fa:16:3e:51:ce:fa", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7138c7dd-49", "ovs_interfaceid": "7138c7dd-4984-4d4f-a631-0d0f531c0044", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.555568] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "5daccb48-b9bf-423a-ab1a-5aa26d8bec87" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.555908] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "5daccb48-b9bf-423a-ab1a-5aa26d8bec87" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.556430] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "5daccb48-b9bf-423a-ab1a-5aa26d8bec87-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.556672] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "5daccb48-b9bf-423a-ab1a-5aa26d8bec87-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.556864] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "5daccb48-b9bf-423a-ab1a-5aa26d8bec87-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.562764] env[61243]: INFO nova.compute.manager [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Terminating instance [ 866.565105] env[61243]: DEBUG nova.compute.manager [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 866.565105] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 866.565734] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbedd013-1f00-4478-b05f-7a5307d8c2ef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.573875] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 866.576699] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd394495-682a-4802-8750-334986b4c921 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.581876] env[61243]: DEBUG nova.network.neutron [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Successfully created port: 5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 866.588604] env[61243]: DEBUG oslo_vmware.api [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 866.588604] env[61243]: value = "task-1338964" [ 866.588604] env[61243]: _type = "Task" [ 866.588604] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.596321] env[61243]: DEBUG oslo_vmware.api [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338964, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.689620] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f44415-7602-738a-7da1-33a48887bd61, 'name': SearchDatastore_Task, 'duration_secs': 0.012504} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.692381] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.692636] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 866.692898] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.693065] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.693268] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 866.693782] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-82ddcb6d-17ba-4ae0-9084-581cfa17d30f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.702799] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 866.703016] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 866.706331] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d518ad52-f59f-4abc-bbeb-97f755d71e15 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.712301] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 866.712301] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52af9a19-dc6d-285a-6d2a-d9143f57e0c1" [ 866.712301] env[61243]: _type = "Task" [ 866.712301] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.719773] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52af9a19-dc6d-285a-6d2a-d9143f57e0c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.822734] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e479b654-29b0-4648-8431-d1f4402b3b48 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.830878] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b8522b-14b3-4cc6-8f4e-57d574d579b1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.860365] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a07562-5041-42bf-83ec-e94ebbd80570 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.868101] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f292d8ba-361e-408d-ac68-793a5fff760f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.881459] env[61243]: DEBUG nova.compute.provider_tree [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.968339] env[61243]: DEBUG oslo_vmware.api [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338963, 'name': ReconfigVM_Task, 'duration_secs': 0.514253} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.968677] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Reconfigured VM instance instance-00000038 to attach disk [datastore2] volume-3111ce1e-7b01-4717-885c-55baf54556d1/volume-3111ce1e-7b01-4717-885c-55baf54556d1.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.973393] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fabccc34-9126-4fd5-8a27-0ef5c15caacf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.989361] env[61243]: DEBUG oslo_vmware.api [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 866.989361] env[61243]: value = "task-1338965" [ 866.989361] env[61243]: _type = "Task" [ 866.989361] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.999704] env[61243]: DEBUG oslo_vmware.api [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338965, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.020461] env[61243]: DEBUG oslo_concurrency.lockutils [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.021140] env[61243]: DEBUG oslo_concurrency.lockutils [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.021303] env[61243]: DEBUG oslo_concurrency.lockutils [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.023537] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9062bcdd-a8f9-44da-b95d-30801118f665 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.042033] env[61243]: DEBUG nova.virt.hardware [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 867.042285] env[61243]: DEBUG nova.virt.hardware [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 867.042462] env[61243]: DEBUG nova.virt.hardware [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 867.042675] env[61243]: DEBUG nova.virt.hardware [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 867.042833] env[61243]: DEBUG nova.virt.hardware [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 867.043009] env[61243]: DEBUG nova.virt.hardware [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 867.043260] env[61243]: DEBUG nova.virt.hardware [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 867.043435] env[61243]: DEBUG nova.virt.hardware [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 867.043608] env[61243]: DEBUG nova.virt.hardware [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 867.043809] env[61243]: DEBUG nova.virt.hardware [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 867.044017] env[61243]: DEBUG nova.virt.hardware [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 867.050631] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Reconfiguring VM to attach interface {{(pid=61243) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 867.051803] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3350253c-6c07-4074-8cf3-992ff667276f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.070818] env[61243]: DEBUG oslo_vmware.api [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 867.070818] env[61243]: value = "task-1338966" [ 867.070818] env[61243]: _type = "Task" [ 867.070818] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.079954] env[61243]: DEBUG oslo_vmware.api [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338966, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.097577] env[61243]: DEBUG oslo_vmware.api [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338964, 'name': PowerOffVM_Task, 'duration_secs': 0.216872} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.097874] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 867.097986] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 867.098259] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-58d12849-9233-4193-ba18-063e835d44e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.170055] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 867.170055] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 867.170055] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Deleting the datastore file [datastore1] 5daccb48-b9bf-423a-ab1a-5aa26d8bec87 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 867.172035] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fa82fa23-eec8-4c5c-95c6-328f900a5586 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.174854] env[61243]: DEBUG nova.compute.manager [req-1c5d1e9e-4dc7-4ba8-9ce9-895c7fca16dd req-0214dcf8-4956-4f11-9904-145e95aecdff service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Received event network-vif-plugged-7138c7dd-4984-4d4f-a631-0d0f531c0044 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.175111] env[61243]: DEBUG oslo_concurrency.lockutils [req-1c5d1e9e-4dc7-4ba8-9ce9-895c7fca16dd req-0214dcf8-4956-4f11-9904-145e95aecdff service nova] Acquiring lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.175675] env[61243]: DEBUG oslo_concurrency.lockutils [req-1c5d1e9e-4dc7-4ba8-9ce9-895c7fca16dd req-0214dcf8-4956-4f11-9904-145e95aecdff service nova] Lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.175874] env[61243]: DEBUG oslo_concurrency.lockutils [req-1c5d1e9e-4dc7-4ba8-9ce9-895c7fca16dd req-0214dcf8-4956-4f11-9904-145e95aecdff service nova] Lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.176108] env[61243]: DEBUG nova.compute.manager [req-1c5d1e9e-4dc7-4ba8-9ce9-895c7fca16dd req-0214dcf8-4956-4f11-9904-145e95aecdff service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] No waiting events found dispatching network-vif-plugged-7138c7dd-4984-4d4f-a631-0d0f531c0044 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 867.176355] env[61243]: WARNING nova.compute.manager [req-1c5d1e9e-4dc7-4ba8-9ce9-895c7fca16dd req-0214dcf8-4956-4f11-9904-145e95aecdff service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Received unexpected event network-vif-plugged-7138c7dd-4984-4d4f-a631-0d0f531c0044 for instance with vm_state active and task_state None. [ 867.176545] env[61243]: DEBUG nova.compute.manager [req-1c5d1e9e-4dc7-4ba8-9ce9-895c7fca16dd req-0214dcf8-4956-4f11-9904-145e95aecdff service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Received event network-changed-7138c7dd-4984-4d4f-a631-0d0f531c0044 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.176729] env[61243]: DEBUG nova.compute.manager [req-1c5d1e9e-4dc7-4ba8-9ce9-895c7fca16dd req-0214dcf8-4956-4f11-9904-145e95aecdff service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Refreshing instance network info cache due to event network-changed-7138c7dd-4984-4d4f-a631-0d0f531c0044. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 867.176941] env[61243]: DEBUG oslo_concurrency.lockutils [req-1c5d1e9e-4dc7-4ba8-9ce9-895c7fca16dd req-0214dcf8-4956-4f11-9904-145e95aecdff service nova] Acquiring lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.177131] env[61243]: DEBUG oslo_concurrency.lockutils [req-1c5d1e9e-4dc7-4ba8-9ce9-895c7fca16dd req-0214dcf8-4956-4f11-9904-145e95aecdff service nova] Acquired lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.177308] env[61243]: DEBUG nova.network.neutron [req-1c5d1e9e-4dc7-4ba8-9ce9-895c7fca16dd req-0214dcf8-4956-4f11-9904-145e95aecdff service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Refreshing network info cache for port 7138c7dd-4984-4d4f-a631-0d0f531c0044 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 867.185671] env[61243]: DEBUG oslo_vmware.api [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 867.185671] env[61243]: value = "task-1338968" [ 867.185671] env[61243]: _type = "Task" [ 867.185671] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.196552] env[61243]: DEBUG oslo_vmware.api [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338968, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.222459] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52af9a19-dc6d-285a-6d2a-d9143f57e0c1, 'name': SearchDatastore_Task, 'duration_secs': 0.020001} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.223256] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9be3e3d2-9599-44df-a50e-05068213e950 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.228957] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 867.228957] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d24f17-c429-ab18-58bc-d2d69c64a185" [ 867.228957] env[61243]: _type = "Task" [ 867.228957] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.236868] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d24f17-c429-ab18-58bc-d2d69c64a185, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.384476] env[61243]: DEBUG nova.scheduler.client.report [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 867.500807] env[61243]: DEBUG oslo_vmware.api [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338965, 'name': ReconfigVM_Task, 'duration_secs': 0.155926} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.501190] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285711', 'volume_id': '3111ce1e-7b01-4717-885c-55baf54556d1', 'name': 'volume-3111ce1e-7b01-4717-885c-55baf54556d1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '07ee984d-476e-484d-ba80-0ec2e411faa9', 'attached_at': '', 'detached_at': '', 'volume_id': '3111ce1e-7b01-4717-885c-55baf54556d1', 'serial': '3111ce1e-7b01-4717-885c-55baf54556d1'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 867.506459] env[61243]: DEBUG nova.compute.manager [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 867.537517] env[61243]: DEBUG nova.virt.hardware [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 867.537801] env[61243]: DEBUG nova.virt.hardware [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 867.537970] env[61243]: DEBUG nova.virt.hardware [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 867.538174] env[61243]: DEBUG nova.virt.hardware [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 867.538336] env[61243]: DEBUG nova.virt.hardware [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 867.538540] env[61243]: DEBUG nova.virt.hardware [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 867.538767] env[61243]: DEBUG nova.virt.hardware [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 867.538975] env[61243]: DEBUG nova.virt.hardware [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 867.539195] env[61243]: DEBUG nova.virt.hardware [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 867.539386] env[61243]: DEBUG nova.virt.hardware [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 867.539679] env[61243]: DEBUG nova.virt.hardware [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 867.540698] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a723dd-1224-4b1b-aa5b-f65f0f563d30 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.549386] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1cb7bc2-bba6-4c39-87f5-81aeadf4aa16 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.580817] env[61243]: DEBUG oslo_vmware.api [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338966, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.675849] env[61243]: DEBUG oslo_vmware.rw_handles [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527d22a4-7427-d947-4e73-f46c6798cdea/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 867.677089] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de1585b8-cf9b-4dc2-86e5-aaa44c769275 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.686217] env[61243]: DEBUG oslo_vmware.rw_handles [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527d22a4-7427-d947-4e73-f46c6798cdea/disk-0.vmdk is in state: ready. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 867.686415] env[61243]: ERROR oslo_vmware.rw_handles [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527d22a4-7427-d947-4e73-f46c6798cdea/disk-0.vmdk due to incomplete transfer. [ 867.686735] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-506a49d1-d7a9-4f12-96fa-0e86160b7f8b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.705420] env[61243]: DEBUG oslo_vmware.api [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1338968, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.203883} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.706622] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.706843] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 867.707050] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 867.707259] env[61243]: INFO nova.compute.manager [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Took 1.14 seconds to destroy the instance on the hypervisor. [ 867.707504] env[61243]: DEBUG oslo.service.loopingcall [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.707738] env[61243]: DEBUG oslo_vmware.rw_handles [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527d22a4-7427-d947-4e73-f46c6798cdea/disk-0.vmdk. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 867.707968] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Uploaded image daf672ed-2d6b-4ef8-8911-3a353c295ac4 to the Glance image server {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 867.711805] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Destroying the VM {{(pid=61243) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 867.715029] env[61243]: DEBUG nova.compute.manager [-] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 867.715029] env[61243]: DEBUG nova.network.neutron [-] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 867.715029] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-200a4eba-2cde-438f-b157-6c7ddde881ff {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.721137] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 867.721137] env[61243]: value = "task-1338969" [ 867.721137] env[61243]: _type = "Task" [ 867.721137] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.730050] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338969, 'name': Destroy_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.741252] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d24f17-c429-ab18-58bc-d2d69c64a185, 'name': SearchDatastore_Task, 'duration_secs': 0.018435} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.741592] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.742033] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] ca230425-aaa7-4c32-81d3-379cbce0d41b/ca230425-aaa7-4c32-81d3-379cbce0d41b.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 867.742387] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d1c021ba-7faa-448b-9e63-6ff1721ea547 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.749074] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 867.749074] env[61243]: value = "task-1338970" [ 867.749074] env[61243]: _type = "Task" [ 867.749074] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.756765] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338970, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.890392] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.890939] env[61243]: DEBUG nova.compute.manager [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 867.893921] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.213s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.895809] env[61243]: INFO nova.compute.claims [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 868.085373] env[61243]: DEBUG oslo_vmware.api [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338966, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.177706] env[61243]: DEBUG nova.network.neutron [req-1c5d1e9e-4dc7-4ba8-9ce9-895c7fca16dd req-0214dcf8-4956-4f11-9904-145e95aecdff service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Updated VIF entry in instance network info cache for port 7138c7dd-4984-4d4f-a631-0d0f531c0044. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 868.177839] env[61243]: DEBUG nova.network.neutron [req-1c5d1e9e-4dc7-4ba8-9ce9-895c7fca16dd req-0214dcf8-4956-4f11-9904-145e95aecdff service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Updating instance_info_cache with network_info: [{"id": "4cd8c409-8daa-4f53-ac5f-d2ac25de247b", "address": "fa:16:3e:11:60:2d", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4cd8c409-8d", "ovs_interfaceid": "4cd8c409-8daa-4f53-ac5f-d2ac25de247b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7138c7dd-4984-4d4f-a631-0d0f531c0044", "address": "fa:16:3e:51:ce:fa", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7138c7dd-49", "ovs_interfaceid": "7138c7dd-4984-4d4f-a631-0d0f531c0044", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.232357] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338969, 'name': Destroy_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.267185] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338970, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.385913] env[61243]: DEBUG nova.network.neutron [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Successfully updated port: 5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 868.400371] env[61243]: DEBUG nova.compute.utils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 868.404255] env[61243]: DEBUG nova.compute.manager [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 868.404255] env[61243]: DEBUG nova.network.neutron [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 868.443162] env[61243]: DEBUG nova.policy [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b9748805e2c48e4a8fd0354e916aa0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '324ed556b6f5419fa0376a7cfc508827', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 868.553540] env[61243]: DEBUG nova.objects.instance [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lazy-loading 'flavor' on Instance uuid 07ee984d-476e-484d-ba80-0ec2e411faa9 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.582691] env[61243]: DEBUG oslo_vmware.api [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338966, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.682146] env[61243]: DEBUG oslo_concurrency.lockutils [req-1c5d1e9e-4dc7-4ba8-9ce9-895c7fca16dd req-0214dcf8-4956-4f11-9904-145e95aecdff service nova] Releasing lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.698695] env[61243]: DEBUG nova.network.neutron [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Successfully created port: a0b324d6-6a73-4956-85c3-38f755003b74 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 868.704316] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fb710d60-4b71-497d-9ab0-1da0235fb1cb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "07ee984d-476e-484d-ba80-0ec2e411faa9" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.731594] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338969, 'name': Destroy_Task, 'duration_secs': 0.682688} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.731825] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Destroyed the VM [ 868.732086] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Deleting Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 868.732353] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-48fabcb1-fa86-4380-aecf-8b5ce4512aec {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.739209] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 868.739209] env[61243]: value = "task-1338971" [ 868.739209] env[61243]: _type = "Task" [ 868.739209] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.746703] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338971, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.759599] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338970, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.565734} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.759599] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] ca230425-aaa7-4c32-81d3-379cbce0d41b/ca230425-aaa7-4c32-81d3-379cbce0d41b.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 868.759599] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 868.759599] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-335f8901-8fed-4dc5-ad4e-1195d43a7a47 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.765427] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 868.765427] env[61243]: value = "task-1338972" [ 868.765427] env[61243]: _type = "Task" [ 868.765427] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.773281] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338972, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.789834] env[61243]: DEBUG nova.network.neutron [-] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.890762] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Acquiring lock "refresh_cache-8df486c4-d1ce-47b0-81a8-ad9e83eaac7f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.890762] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Acquired lock "refresh_cache-8df486c4-d1ce-47b0-81a8-ad9e83eaac7f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.890762] env[61243]: DEBUG nova.network.neutron [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 868.904807] env[61243]: DEBUG nova.compute.manager [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 869.059486] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c364df7c-5f0c-483b-8da9-4b1e9db1dc82 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.807s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.060948] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fb710d60-4b71-497d-9ab0-1da0235fb1cb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.356s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.060948] env[61243]: DEBUG nova.compute.manager [None req-fb710d60-4b71-497d-9ab0-1da0235fb1cb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.061754] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-564b8a17-190d-474c-9d39-ed4552faddf1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.072057] env[61243]: DEBUG nova.compute.manager [None req-fb710d60-4b71-497d-9ab0-1da0235fb1cb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61243) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 869.072687] env[61243]: DEBUG nova.objects.instance [None req-fb710d60-4b71-497d-9ab0-1da0235fb1cb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lazy-loading 'flavor' on Instance uuid 07ee984d-476e-484d-ba80-0ec2e411faa9 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.085943] env[61243]: DEBUG oslo_vmware.api [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338966, 'name': ReconfigVM_Task, 'duration_secs': 1.681005} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.086448] env[61243]: DEBUG oslo_concurrency.lockutils [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.086675] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Reconfigured VM to attach interface {{(pid=61243) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 869.199671] env[61243]: DEBUG nova.compute.manager [req-c2c065ac-9022-457e-9ac3-12056543bc76 req-3722a1c2-5455-4f6f-bffc-31ec40808094 service nova] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Received event network-vif-plugged-5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.199902] env[61243]: DEBUG oslo_concurrency.lockutils [req-c2c065ac-9022-457e-9ac3-12056543bc76 req-3722a1c2-5455-4f6f-bffc-31ec40808094 service nova] Acquiring lock "8df486c4-d1ce-47b0-81a8-ad9e83eaac7f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.200133] env[61243]: DEBUG oslo_concurrency.lockutils [req-c2c065ac-9022-457e-9ac3-12056543bc76 req-3722a1c2-5455-4f6f-bffc-31ec40808094 service nova] Lock "8df486c4-d1ce-47b0-81a8-ad9e83eaac7f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.200310] env[61243]: DEBUG oslo_concurrency.lockutils [req-c2c065ac-9022-457e-9ac3-12056543bc76 req-3722a1c2-5455-4f6f-bffc-31ec40808094 service nova] Lock "8df486c4-d1ce-47b0-81a8-ad9e83eaac7f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.200479] env[61243]: DEBUG nova.compute.manager [req-c2c065ac-9022-457e-9ac3-12056543bc76 req-3722a1c2-5455-4f6f-bffc-31ec40808094 service nova] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] No waiting events found dispatching network-vif-plugged-5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 869.200649] env[61243]: WARNING nova.compute.manager [req-c2c065ac-9022-457e-9ac3-12056543bc76 req-3722a1c2-5455-4f6f-bffc-31ec40808094 service nova] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Received unexpected event network-vif-plugged-5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7 for instance with vm_state building and task_state spawning. [ 869.200816] env[61243]: DEBUG nova.compute.manager [req-c2c065ac-9022-457e-9ac3-12056543bc76 req-3722a1c2-5455-4f6f-bffc-31ec40808094 service nova] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Received event network-vif-deleted-1aa2d7ff-eac6-472b-b7a7-57c542d3f111 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.200991] env[61243]: DEBUG nova.compute.manager [req-c2c065ac-9022-457e-9ac3-12056543bc76 req-3722a1c2-5455-4f6f-bffc-31ec40808094 service nova] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Received event network-changed-5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.201168] env[61243]: DEBUG nova.compute.manager [req-c2c065ac-9022-457e-9ac3-12056543bc76 req-3722a1c2-5455-4f6f-bffc-31ec40808094 service nova] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Refreshing instance network info cache due to event network-changed-5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 869.201351] env[61243]: DEBUG oslo_concurrency.lockutils [req-c2c065ac-9022-457e-9ac3-12056543bc76 req-3722a1c2-5455-4f6f-bffc-31ec40808094 service nova] Acquiring lock "refresh_cache-8df486c4-d1ce-47b0-81a8-ad9e83eaac7f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.206465] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e3460f-8b47-4608-ae60-21542a1ab6cf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.213720] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0478446-c685-41a8-8dd5-049e20a0fe6f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.247532] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d2eddab-7e8f-4e85-9371-db3250335db9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.257183] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a632da5-a91f-4af0-9723-bd414c33f903 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.260694] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338971, 'name': RemoveSnapshot_Task, 'duration_secs': 0.406162} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.260949] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Deleted Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 869.261233] env[61243]: DEBUG nova.compute.manager [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.262288] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52e0e9e0-f2c4-40d8-8624-bcdfa8900101 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.272058] env[61243]: DEBUG nova.compute.provider_tree [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 869.284154] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338972, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.21312} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.284399] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 869.285123] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc50294-922f-430e-9141-95dcf349c5ca {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.297907] env[61243]: INFO nova.compute.manager [-] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Took 1.59 seconds to deallocate network for instance. [ 869.307220] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] ca230425-aaa7-4c32-81d3-379cbce0d41b/ca230425-aaa7-4c32-81d3-379cbce0d41b.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 869.308461] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-407809ca-1d71-412c-b35d-c2075eb55c67 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.330134] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 869.330134] env[61243]: value = "task-1338973" [ 869.330134] env[61243]: _type = "Task" [ 869.330134] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.337737] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338973, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.423528] env[61243]: DEBUG nova.network.neutron [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 869.576732] env[61243]: DEBUG nova.network.neutron [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Updating instance_info_cache with network_info: [{"id": "5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7", "address": "fa:16:3e:90:9c:10", "network": {"id": "b77f74a5-6460-46f1-910d-b64eaca5e69f", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1175228587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf51131cc0e247669ddf5495ffebd8d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cebc48c-6a77-46bf-9c12-ac130e4d7d76", "external-id": "nsx-vlan-transportzone-382", "segmentation_id": 382, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ae039d5-ce", "ovs_interfaceid": "5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.586141] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb710d60-4b71-497d-9ab0-1da0235fb1cb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 869.586212] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-574c7886-5187-49b3-9579-f7a8e221b96f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.591259] env[61243]: DEBUG oslo_concurrency.lockutils [None req-82400ff4-da29-4340-b3dd-a7701453d8ef tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.160s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.593830] env[61243]: DEBUG oslo_vmware.api [None req-fb710d60-4b71-497d-9ab0-1da0235fb1cb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 869.593830] env[61243]: value = "task-1338974" [ 869.593830] env[61243]: _type = "Task" [ 869.593830] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.603608] env[61243]: DEBUG oslo_vmware.api [None req-fb710d60-4b71-497d-9ab0-1da0235fb1cb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338974, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.783712] env[61243]: INFO nova.compute.manager [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Shelve offloading [ 869.786393] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 869.786742] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc5f0a92-e377-49be-8ee3-856218a933c9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.793925] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 869.793925] env[61243]: value = "task-1338975" [ 869.793925] env[61243]: _type = "Task" [ 869.793925] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.802703] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338975, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.807587] env[61243]: DEBUG nova.scheduler.client.report [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 92 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 869.807846] env[61243]: DEBUG nova.compute.provider_tree [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 92 to 93 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 869.808045] env[61243]: DEBUG nova.compute.provider_tree [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 869.826082] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.840391] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338973, 'name': ReconfigVM_Task, 'duration_secs': 0.314959} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.840678] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Reconfigured VM instance instance-00000049 to attach disk [datastore2] ca230425-aaa7-4c32-81d3-379cbce0d41b/ca230425-aaa7-4c32-81d3-379cbce0d41b.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 869.841327] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b320bcd-9531-4197-abd0-8bb47c237cbd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.847943] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 869.847943] env[61243]: value = "task-1338976" [ 869.847943] env[61243]: _type = "Task" [ 869.847943] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.855748] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338976, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.916204] env[61243]: DEBUG nova.compute.manager [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 869.941685] env[61243]: DEBUG nova.virt.hardware [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.942028] env[61243]: DEBUG nova.virt.hardware [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.942122] env[61243]: DEBUG nova.virt.hardware [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.942314] env[61243]: DEBUG nova.virt.hardware [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.942468] env[61243]: DEBUG nova.virt.hardware [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.942619] env[61243]: DEBUG nova.virt.hardware [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.942826] env[61243]: DEBUG nova.virt.hardware [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.942988] env[61243]: DEBUG nova.virt.hardware [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.943176] env[61243]: DEBUG nova.virt.hardware [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.943363] env[61243]: DEBUG nova.virt.hardware [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.943636] env[61243]: DEBUG nova.virt.hardware [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.944571] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb38e0b0-b0e5-4785-a1ae-4adf324acddd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.954358] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-270a9900-eebc-473b-bd74-e1571fe12048 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.079387] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Releasing lock "refresh_cache-8df486c4-d1ce-47b0-81a8-ad9e83eaac7f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.079745] env[61243]: DEBUG nova.compute.manager [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Instance network_info: |[{"id": "5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7", "address": "fa:16:3e:90:9c:10", "network": {"id": "b77f74a5-6460-46f1-910d-b64eaca5e69f", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1175228587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf51131cc0e247669ddf5495ffebd8d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cebc48c-6a77-46bf-9c12-ac130e4d7d76", "external-id": "nsx-vlan-transportzone-382", "segmentation_id": 382, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ae039d5-ce", "ovs_interfaceid": "5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 870.080079] env[61243]: DEBUG oslo_concurrency.lockutils [req-c2c065ac-9022-457e-9ac3-12056543bc76 req-3722a1c2-5455-4f6f-bffc-31ec40808094 service nova] Acquired lock "refresh_cache-8df486c4-d1ce-47b0-81a8-ad9e83eaac7f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.080287] env[61243]: DEBUG nova.network.neutron [req-c2c065ac-9022-457e-9ac3-12056543bc76 req-3722a1c2-5455-4f6f-bffc-31ec40808094 service nova] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Refreshing network info cache for port 5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 870.081676] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:9c:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cebc48c-6a77-46bf-9c12-ac130e4d7d76', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.090053] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Creating folder: Project (bf51131cc0e247669ddf5495ffebd8d5). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 870.092676] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce6d44c9-55bb-47a2-90e5-64d659465140 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.104206] env[61243]: DEBUG oslo_vmware.api [None req-fb710d60-4b71-497d-9ab0-1da0235fb1cb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338974, 'name': PowerOffVM_Task, 'duration_secs': 0.237231} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.104455] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb710d60-4b71-497d-9ab0-1da0235fb1cb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 870.104633] env[61243]: DEBUG nova.compute.manager [None req-fb710d60-4b71-497d-9ab0-1da0235fb1cb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.106300] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d94778a-8737-44a1-a9bb-57319f05ee5e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.108561] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Created folder: Project (bf51131cc0e247669ddf5495ffebd8d5) in parent group-v285636. [ 870.108744] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Creating folder: Instances. Parent ref: group-v285715. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 870.108960] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5bc64354-338a-4f00-a9e9-03fdf0dcb49c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.117380] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Created folder: Instances in parent group-v285715. [ 870.117608] env[61243]: DEBUG oslo.service.loopingcall [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.117965] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 870.118200] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6f0ffcc2-bf30-4cf8-8c46-ae8d78266c47 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.137341] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.137341] env[61243]: value = "task-1338979" [ 870.137341] env[61243]: _type = "Task" [ 870.137341] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.144918] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338979, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.304355] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] VM already powered off {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 870.304744] env[61243]: DEBUG nova.compute.manager [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.305450] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d01a6790-b363-416a-98bc-0f5808651ca5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.312698] env[61243]: DEBUG nova.network.neutron [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Successfully updated port: a0b324d6-6a73-4956-85c3-38f755003b74 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 870.314232] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.420s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.314703] env[61243]: DEBUG nova.compute.manager [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 870.317762] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.389s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.317944] env[61243]: DEBUG nova.objects.instance [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61243) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 870.322980] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "refresh_cache-ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.323165] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "refresh_cache-ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.323517] env[61243]: DEBUG nova.network.neutron [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 870.358428] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338976, 'name': Rename_Task, 'duration_secs': 0.151685} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.358702] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 870.358949] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2fece2e-fb16-443c-9eb8-e2b016af0898 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.365817] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 870.365817] env[61243]: value = "task-1338980" [ 870.365817] env[61243]: _type = "Task" [ 870.365817] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.373523] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338980, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.374333] env[61243]: DEBUG nova.network.neutron [req-c2c065ac-9022-457e-9ac3-12056543bc76 req-3722a1c2-5455-4f6f-bffc-31ec40808094 service nova] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Updated VIF entry in instance network info cache for port 5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 870.374683] env[61243]: DEBUG nova.network.neutron [req-c2c065ac-9022-457e-9ac3-12056543bc76 req-3722a1c2-5455-4f6f-bffc-31ec40808094 service nova] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Updating instance_info_cache with network_info: [{"id": "5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7", "address": "fa:16:3e:90:9c:10", "network": {"id": "b77f74a5-6460-46f1-910d-b64eaca5e69f", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1175228587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf51131cc0e247669ddf5495ffebd8d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cebc48c-6a77-46bf-9c12-ac130e4d7d76", "external-id": "nsx-vlan-transportzone-382", "segmentation_id": 382, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ae039d5-ce", "ovs_interfaceid": "5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.623985] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fb710d60-4b71-497d-9ab0-1da0235fb1cb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.563s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.648740] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338979, 'name': CreateVM_Task, 'duration_secs': 0.381248} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.648948] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 870.649715] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.649924] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.650303] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 870.650613] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8df95bd5-0008-4f0c-9ed4-c2993a6abb6e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.655103] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Waiting for the task: (returnval){ [ 870.655103] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52903a55-6dfa-58f3-bde0-1dde04f55fc1" [ 870.655103] env[61243]: _type = "Task" [ 870.655103] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.663173] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52903a55-6dfa-58f3-bde0-1dde04f55fc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.818680] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "refresh_cache-3f839aa5-2b9a-4807-b63b-931f74455532" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.819121] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "refresh_cache-3f839aa5-2b9a-4807-b63b-931f74455532" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.819121] env[61243]: DEBUG nova.network.neutron [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 870.825320] env[61243]: DEBUG nova.compute.utils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 870.832127] env[61243]: DEBUG nova.compute.manager [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 870.832310] env[61243]: DEBUG nova.network.neutron [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 870.876047] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338980, 'name': PowerOnVM_Task} progress is 91%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.876562] env[61243]: DEBUG oslo_concurrency.lockutils [req-c2c065ac-9022-457e-9ac3-12056543bc76 req-3722a1c2-5455-4f6f-bffc-31ec40808094 service nova] Releasing lock "refresh_cache-8df486c4-d1ce-47b0-81a8-ad9e83eaac7f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.905334] env[61243]: DEBUG nova.policy [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cc77f359ba44dc0bc8e053d5a4b63ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3248cca8a8aa4cad84b8bd5dcf7ed0c3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 871.165904] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52903a55-6dfa-58f3-bde0-1dde04f55fc1, 'name': SearchDatastore_Task, 'duration_secs': 0.015865} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.166311] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.166576] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 871.166826] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.166983] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.167238] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 871.167537] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6f4f9624-cb21-4bd2-94df-fdeb06de433d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.191320] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 871.191676] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 871.192569] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d50f8a7b-aff1-4cb5-802d-fdb20a8d16b4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.197967] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Waiting for the task: (returnval){ [ 871.197967] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525c1e1f-3575-137f-bd21-8021dfdf0487" [ 871.197967] env[61243]: _type = "Task" [ 871.197967] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.209158] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525c1e1f-3575-137f-bd21-8021dfdf0487, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.223605] env[61243]: DEBUG nova.compute.manager [req-c89d4536-fce9-414d-9ecf-cb1e608df3f2 req-722af606-59b9-4ef9-8651-9931af0f8889 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Received event network-vif-plugged-a0b324d6-6a73-4956-85c3-38f755003b74 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.223821] env[61243]: DEBUG oslo_concurrency.lockutils [req-c89d4536-fce9-414d-9ecf-cb1e608df3f2 req-722af606-59b9-4ef9-8651-9931af0f8889 service nova] Acquiring lock "3f839aa5-2b9a-4807-b63b-931f74455532-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.224041] env[61243]: DEBUG oslo_concurrency.lockutils [req-c89d4536-fce9-414d-9ecf-cb1e608df3f2 req-722af606-59b9-4ef9-8651-9931af0f8889 service nova] Lock "3f839aa5-2b9a-4807-b63b-931f74455532-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.224246] env[61243]: DEBUG oslo_concurrency.lockutils [req-c89d4536-fce9-414d-9ecf-cb1e608df3f2 req-722af606-59b9-4ef9-8651-9931af0f8889 service nova] Lock "3f839aa5-2b9a-4807-b63b-931f74455532-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.224431] env[61243]: DEBUG nova.compute.manager [req-c89d4536-fce9-414d-9ecf-cb1e608df3f2 req-722af606-59b9-4ef9-8651-9931af0f8889 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] No waiting events found dispatching network-vif-plugged-a0b324d6-6a73-4956-85c3-38f755003b74 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 871.224607] env[61243]: WARNING nova.compute.manager [req-c89d4536-fce9-414d-9ecf-cb1e608df3f2 req-722af606-59b9-4ef9-8651-9931af0f8889 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Received unexpected event network-vif-plugged-a0b324d6-6a73-4956-85c3-38f755003b74 for instance with vm_state building and task_state spawning. [ 871.224769] env[61243]: DEBUG nova.compute.manager [req-c89d4536-fce9-414d-9ecf-cb1e608df3f2 req-722af606-59b9-4ef9-8651-9931af0f8889 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Received event network-changed-a0b324d6-6a73-4956-85c3-38f755003b74 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.224928] env[61243]: DEBUG nova.compute.manager [req-c89d4536-fce9-414d-9ecf-cb1e608df3f2 req-722af606-59b9-4ef9-8651-9931af0f8889 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Refreshing instance network info cache due to event network-changed-a0b324d6-6a73-4956-85c3-38f755003b74. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 871.225935] env[61243]: DEBUG oslo_concurrency.lockutils [req-c89d4536-fce9-414d-9ecf-cb1e608df3f2 req-722af606-59b9-4ef9-8651-9931af0f8889 service nova] Acquiring lock "refresh_cache-3f839aa5-2b9a-4807-b63b-931f74455532" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.325420] env[61243]: DEBUG nova.objects.instance [None req-c5424924-7f10-4bcc-ba94-a5542aaab3ff tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lazy-loading 'flavor' on Instance uuid 07ee984d-476e-484d-ba80-0ec2e411faa9 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.336149] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c4c9655-7052-4f0b-ac27-e8c7d43b04f2 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.337405] env[61243]: DEBUG nova.compute.manager [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 871.339955] env[61243]: DEBUG oslo_concurrency.lockutils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.649s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.344204] env[61243]: INFO nova.compute.claims [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 871.362490] env[61243]: DEBUG nova.network.neutron [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Updating instance_info_cache with network_info: [{"id": "5ea2a618-21af-437b-b35b-69f2a38cc3dc", "address": "fa:16:3e:6f:f0:da", "network": {"id": "0788f9ac-5a15-4431-a711-9b4f6cf79fbc", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1210957657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "113713170fb94b8a8d7985b9a02faf77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c7821ea-f92f-4f06-a4cb-05e1186a9d22", "external-id": "nsx-vlan-transportzone-69", "segmentation_id": 69, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ea2a618-21", "ovs_interfaceid": "5ea2a618-21af-437b-b35b-69f2a38cc3dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.377622] env[61243]: DEBUG oslo_vmware.api [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338980, 'name': PowerOnVM_Task, 'duration_secs': 0.579154} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.378475] env[61243]: DEBUG nova.network.neutron [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 871.380305] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 871.380519] env[61243]: INFO nova.compute.manager [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Took 10.28 seconds to spawn the instance on the hypervisor. [ 871.380704] env[61243]: DEBUG nova.compute.manager [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 871.381960] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37521913-5f63-4ddf-8d3b-4e1a1e9f723e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.541357] env[61243]: DEBUG nova.network.neutron [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Successfully created port: c706f966-e43c-4d65-b536-f61daa91c8ef {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 871.575197] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "interface-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00-7138c7dd-4984-4d4f-a631-0d0f531c0044" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.575523] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00-7138c7dd-4984-4d4f-a631-0d0f531c0044" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.603534] env[61243]: DEBUG nova.network.neutron [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Updating instance_info_cache with network_info: [{"id": "a0b324d6-6a73-4956-85c3-38f755003b74", "address": "fa:16:3e:16:b3:46", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0b324d6-6a", "ovs_interfaceid": "a0b324d6-6a73-4956-85c3-38f755003b74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.711567] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525c1e1f-3575-137f-bd21-8021dfdf0487, 'name': SearchDatastore_Task, 'duration_secs': 0.023394} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.712395] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4aaf1cf-e131-4bd8-b5c4-5ca2a8caa331 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.718127] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Waiting for the task: (returnval){ [ 871.718127] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e09915-c250-3cce-1d30-5a7f420db881" [ 871.718127] env[61243]: _type = "Task" [ 871.718127] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.725717] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e09915-c250-3cce-1d30-5a7f420db881, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.833971] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c5424924-7f10-4bcc-ba94-a5542aaab3ff tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.833971] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c5424924-7f10-4bcc-ba94-a5542aaab3ff tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquired lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.834105] env[61243]: DEBUG nova.network.neutron [None req-c5424924-7f10-4bcc-ba94-a5542aaab3ff tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.834271] env[61243]: DEBUG nova.objects.instance [None req-c5424924-7f10-4bcc-ba94-a5542aaab3ff tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lazy-loading 'info_cache' on Instance uuid 07ee984d-476e-484d-ba80-0ec2e411faa9 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.865841] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "refresh_cache-ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.902027] env[61243]: INFO nova.compute.manager [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Took 28.36 seconds to build instance. [ 872.080029] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.080029] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.081055] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351d99a9-3dbb-4577-938f-30352e170fee {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.101997] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c53fbf98-6722-4188-b0a5-a65485ba46e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.106494] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "refresh_cache-3f839aa5-2b9a-4807-b63b-931f74455532" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.106823] env[61243]: DEBUG nova.compute.manager [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Instance network_info: |[{"id": "a0b324d6-6a73-4956-85c3-38f755003b74", "address": "fa:16:3e:16:b3:46", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0b324d6-6a", "ovs_interfaceid": "a0b324d6-6a73-4956-85c3-38f755003b74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 872.129979] env[61243]: DEBUG oslo_concurrency.lockutils [req-c89d4536-fce9-414d-9ecf-cb1e608df3f2 req-722af606-59b9-4ef9-8651-9931af0f8889 service nova] Acquired lock "refresh_cache-3f839aa5-2b9a-4807-b63b-931f74455532" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.133126] env[61243]: DEBUG nova.network.neutron [req-c89d4536-fce9-414d-9ecf-cb1e608df3f2 req-722af606-59b9-4ef9-8651-9931af0f8889 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Refreshing network info cache for port a0b324d6-6a73-4956-85c3-38f755003b74 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 872.133126] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:b3:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ff3ecd2f-0b10-4faf-a512-fd7a20c28df1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a0b324d6-6a73-4956-85c3-38f755003b74', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.139236] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Creating folder: Project (324ed556b6f5419fa0376a7cfc508827). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.144933] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Reconfiguring VM to detach interface {{(pid=61243) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 872.146410] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aafd9cb9-0bda-4e22-a7cf-40d3d19e7105 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.148130] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96440f22-cc6b-487f-b9e0-2be752da508c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.163789] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 872.164618] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c7b2e3b-b102-4ca5-ba57-7424da161bab {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.168851] env[61243]: DEBUG oslo_vmware.api [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 872.168851] env[61243]: value = "task-1338982" [ 872.168851] env[61243]: _type = "Task" [ 872.168851] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.173206] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Created folder: Project (324ed556b6f5419fa0376a7cfc508827) in parent group-v285636. [ 872.173398] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Creating folder: Instances. Parent ref: group-v285718. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.175586] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c10c75b2-5a3b-4052-96bd-497f8ca6db6a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.177057] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 872.180285] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c28fb461-ff57-4772-b0e5-ab724fcac4e4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.181669] env[61243]: DEBUG oslo_vmware.api [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338982, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.191564] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Created folder: Instances in parent group-v285718. [ 872.191812] env[61243]: DEBUG oslo.service.loopingcall [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.192050] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 872.192228] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d9446486-038a-4fd1-abda-6f799cd02309 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.212621] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.212621] env[61243]: value = "task-1338985" [ 872.212621] env[61243]: _type = "Task" [ 872.212621] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.220429] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338985, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.228604] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e09915-c250-3cce-1d30-5a7f420db881, 'name': SearchDatastore_Task, 'duration_secs': 0.009415} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.228871] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.229154] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f/8df486c4-d1ce-47b0-81a8-ad9e83eaac7f.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 872.229414] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6515aa7e-5e1f-4f8b-b8c3-2f00fd86adf7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.235360] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Waiting for the task: (returnval){ [ 872.235360] env[61243]: value = "task-1338986" [ 872.235360] env[61243]: _type = "Task" [ 872.235360] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.245455] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1338986, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.248686] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 872.249067] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 872.249288] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Deleting the datastore file [datastore2] ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.249597] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78518a28-a324-465f-bdb0-40719cacf8c6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.256836] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 872.256836] env[61243]: value = "task-1338987" [ 872.256836] env[61243]: _type = "Task" [ 872.256836] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.264417] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338987, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.337885] env[61243]: DEBUG nova.objects.base [None req-c5424924-7f10-4bcc-ba94-a5542aaab3ff tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Object Instance<07ee984d-476e-484d-ba80-0ec2e411faa9> lazy-loaded attributes: flavor,info_cache {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 872.352493] env[61243]: DEBUG nova.compute.manager [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 872.385115] env[61243]: DEBUG nova.virt.hardware [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 872.385382] env[61243]: DEBUG nova.virt.hardware [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 872.386211] env[61243]: DEBUG nova.virt.hardware [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 872.386211] env[61243]: DEBUG nova.virt.hardware [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 872.386211] env[61243]: DEBUG nova.virt.hardware [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 872.386365] env[61243]: DEBUG nova.virt.hardware [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 872.386592] env[61243]: DEBUG nova.virt.hardware [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 872.386697] env[61243]: DEBUG nova.virt.hardware [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 872.386873] env[61243]: DEBUG nova.virt.hardware [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 872.387048] env[61243]: DEBUG nova.virt.hardware [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 872.387284] env[61243]: DEBUG nova.virt.hardware [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 872.388555] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a2fbb5-5aa4-4d53-aaf6-e4b755fb74c9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.399672] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-814889b5-318d-44d7-b00d-18157787423a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.403966] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dde73a26-8db1-4193-a434-69028ad8f19b tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "ca230425-aaa7-4c32-81d3-379cbce0d41b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.019s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.676513] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c773ed05-4e92-452d-b9eb-7e9cf436fd58 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.683898] env[61243]: DEBUG oslo_vmware.api [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338982, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.691148] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e070d93-4fec-4b40-b5f3-2edc4b0a3e43 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.730869] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a35591d-cf3e-4268-b45f-a2bee581e9a4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.743856] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338985, 'name': CreateVM_Task, 'duration_secs': 0.334629} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.746502] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 872.747494] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.747698] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.748143] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 872.749510] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-218ca28f-41b2-482b-9218-2a0740d025e5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.761731] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1781bfd6-ff71-4503-9e3d-1984b08a8787 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.763459] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1338986, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.777889] env[61243]: DEBUG nova.compute.provider_tree [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 872.782876] env[61243]: DEBUG oslo_vmware.api [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1338987, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179221} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.783105] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 872.783105] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5291930e-8e54-fb1d-d97c-a77f926006aa" [ 872.783105] env[61243]: _type = "Task" [ 872.783105] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.783573] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 872.783770] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 872.783950] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 872.796774] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5291930e-8e54-fb1d-d97c-a77f926006aa, 'name': SearchDatastore_Task, 'duration_secs': 0.009133} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.797257] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.797337] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 872.797862] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.797862] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.797984] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 872.798766] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be2b7c84-2c6a-4711-b31a-aeecf48f98ad {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.805568] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 872.806248] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 872.807115] env[61243]: INFO nova.scheduler.client.report [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Deleted allocations for instance ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5 [ 872.809435] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d516afd1-81a8-46d4-894d-79bc9798abed {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.815307] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 872.815307] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528f307d-bf07-e6e4-ca2f-31d2f7259d7e" [ 872.815307] env[61243]: _type = "Task" [ 872.815307] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.823529] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528f307d-bf07-e6e4-ca2f-31d2f7259d7e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.961421] env[61243]: DEBUG oslo_concurrency.lockutils [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "ca230425-aaa7-4c32-81d3-379cbce0d41b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.961758] env[61243]: DEBUG oslo_concurrency.lockutils [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "ca230425-aaa7-4c32-81d3-379cbce0d41b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.962027] env[61243]: DEBUG oslo_concurrency.lockutils [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "ca230425-aaa7-4c32-81d3-379cbce0d41b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.962189] env[61243]: DEBUG oslo_concurrency.lockutils [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "ca230425-aaa7-4c32-81d3-379cbce0d41b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.962363] env[61243]: DEBUG oslo_concurrency.lockutils [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "ca230425-aaa7-4c32-81d3-379cbce0d41b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.964435] env[61243]: INFO nova.compute.manager [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Terminating instance [ 872.966282] env[61243]: DEBUG nova.compute.manager [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 872.966477] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 872.967425] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d589b3f5-ded3-4077-8de3-99e884e85eb7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.974829] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 872.975086] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b247445d-428c-4934-84ef-f55fa63d06af {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.981726] env[61243]: DEBUG oslo_vmware.api [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 872.981726] env[61243]: value = "task-1338988" [ 872.981726] env[61243]: _type = "Task" [ 872.981726] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.989780] env[61243]: DEBUG oslo_vmware.api [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338988, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.009173] env[61243]: DEBUG nova.network.neutron [req-c89d4536-fce9-414d-9ecf-cb1e608df3f2 req-722af606-59b9-4ef9-8651-9931af0f8889 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Updated VIF entry in instance network info cache for port a0b324d6-6a73-4956-85c3-38f755003b74. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 873.010083] env[61243]: DEBUG nova.network.neutron [req-c89d4536-fce9-414d-9ecf-cb1e608df3f2 req-722af606-59b9-4ef9-8651-9931af0f8889 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Updating instance_info_cache with network_info: [{"id": "a0b324d6-6a73-4956-85c3-38f755003b74", "address": "fa:16:3e:16:b3:46", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0b324d6-6a", "ovs_interfaceid": "a0b324d6-6a73-4956-85c3-38f755003b74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.180138] env[61243]: DEBUG oslo_vmware.api [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338982, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.233743] env[61243]: DEBUG nova.network.neutron [None req-c5424924-7f10-4bcc-ba94-a5542aaab3ff tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Updating instance_info_cache with network_info: [{"id": "af7195a1-9a59-4b3a-b62d-7e907c94ce7b", "address": "fa:16:3e:ee:b7:e1", "network": {"id": "c6d4943d-4a8d-41b4-8023-c4f1549d312d", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1045259606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8549626a07e941e1ab03452355a0f116", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf7195a1-9a", "ovs_interfaceid": "af7195a1-9a59-4b3a-b62d-7e907c94ce7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.247013] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1338986, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533759} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.247339] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f/8df486c4-d1ce-47b0-81a8-ad9e83eaac7f.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 873.247842] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 873.247842] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-39aa86d8-935b-4d4c-a278-8fe31f405c28 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.255285] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Waiting for the task: (returnval){ [ 873.255285] env[61243]: value = "task-1338989" [ 873.255285] env[61243]: _type = "Task" [ 873.255285] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.260815] env[61243]: DEBUG nova.compute.manager [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Received event network-vif-unplugged-5ea2a618-21af-437b-b35b-69f2a38cc3dc {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.261053] env[61243]: DEBUG oslo_concurrency.lockutils [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] Acquiring lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.261331] env[61243]: DEBUG oslo_concurrency.lockutils [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] Lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.261465] env[61243]: DEBUG oslo_concurrency.lockutils [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] Lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.261637] env[61243]: DEBUG nova.compute.manager [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] No waiting events found dispatching network-vif-unplugged-5ea2a618-21af-437b-b35b-69f2a38cc3dc {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 873.261811] env[61243]: WARNING nova.compute.manager [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Received unexpected event network-vif-unplugged-5ea2a618-21af-437b-b35b-69f2a38cc3dc for instance with vm_state shelved_offloaded and task_state None. [ 873.261975] env[61243]: DEBUG nova.compute.manager [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Received event network-changed-5ea2a618-21af-437b-b35b-69f2a38cc3dc {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.262146] env[61243]: DEBUG nova.compute.manager [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Refreshing instance network info cache due to event network-changed-5ea2a618-21af-437b-b35b-69f2a38cc3dc. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 873.262338] env[61243]: DEBUG oslo_concurrency.lockutils [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] Acquiring lock "refresh_cache-ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.262477] env[61243]: DEBUG oslo_concurrency.lockutils [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] Acquired lock "refresh_cache-ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.262637] env[61243]: DEBUG nova.network.neutron [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Refreshing network info cache for port 5ea2a618-21af-437b-b35b-69f2a38cc3dc {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 873.270320] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1338989, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.302932] env[61243]: DEBUG nova.network.neutron [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Successfully updated port: c706f966-e43c-4d65-b536-f61daa91c8ef {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 873.314108] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.325123] env[61243]: DEBUG nova.scheduler.client.report [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 93 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 873.325408] env[61243]: DEBUG nova.compute.provider_tree [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 93 to 94 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 873.325598] env[61243]: DEBUG nova.compute.provider_tree [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 873.335676] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528f307d-bf07-e6e4-ca2f-31d2f7259d7e, 'name': SearchDatastore_Task, 'duration_secs': 0.009486} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.336685] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4e96d2f-36c3-4437-ab35-7c048c010806 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.342489] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 873.342489] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52bb3d26-25da-4616-b437-1e33a81f41fb" [ 873.342489] env[61243]: _type = "Task" [ 873.342489] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.350491] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52bb3d26-25da-4616-b437-1e33a81f41fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.491573] env[61243]: DEBUG oslo_vmware.api [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338988, 'name': PowerOffVM_Task, 'duration_secs': 0.203344} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.491866] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 873.493075] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 873.493075] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b4237692-bc1c-4a87-816d-2b32502cef22 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.512906] env[61243]: DEBUG oslo_concurrency.lockutils [req-c89d4536-fce9-414d-9ecf-cb1e608df3f2 req-722af606-59b9-4ef9-8651-9931af0f8889 service nova] Releasing lock "refresh_cache-3f839aa5-2b9a-4807-b63b-931f74455532" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.681293] env[61243]: DEBUG oslo_vmware.api [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338982, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.734165] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c5424924-7f10-4bcc-ba94-a5542aaab3ff tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Releasing lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.762728] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 873.762956] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 873.763167] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleting the datastore file [datastore2] ca230425-aaa7-4c32-81d3-379cbce0d41b {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.766315] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1011ce9-cbec-4d4f-a43f-7c8e324d7ddf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.769745] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1338989, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067886} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.769996] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 873.771305] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c34eceb-40ff-4137-bb68-ad3d1591e4f9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.774525] env[61243]: DEBUG oslo_vmware.api [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 873.774525] env[61243]: value = "task-1338991" [ 873.774525] env[61243]: _type = "Task" [ 873.774525] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.796776] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f/8df486c4-d1ce-47b0-81a8-ad9e83eaac7f.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.797400] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fa65b47-d652-450b-b272-fca424d3d4ec {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.817687] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "refresh_cache-783d861c-8930-406e-8986-ef995de4c9dd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.817861] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired lock "refresh_cache-783d861c-8930-406e-8986-ef995de4c9dd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.818248] env[61243]: DEBUG nova.network.neutron [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 873.819177] env[61243]: DEBUG oslo_vmware.api [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338991, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.825543] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Waiting for the task: (returnval){ [ 873.825543] env[61243]: value = "task-1338992" [ 873.825543] env[61243]: _type = "Task" [ 873.825543] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.830306] env[61243]: DEBUG oslo_concurrency.lockutils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.830855] env[61243]: DEBUG nova.compute.manager [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 873.833879] env[61243]: DEBUG oslo_concurrency.lockutils [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.137s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.834118] env[61243]: DEBUG nova.objects.instance [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lazy-loading 'resources' on Instance uuid 9a66fd05-1880-423a-8f60-4f11f1c75ab3 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.839316] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1338992, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.852614] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52bb3d26-25da-4616-b437-1e33a81f41fb, 'name': SearchDatastore_Task, 'duration_secs': 0.008977} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.852905] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.853211] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 3f839aa5-2b9a-4807-b63b-931f74455532/3f839aa5-2b9a-4807-b63b-931f74455532.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 873.853510] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e4eca7c8-a254-4b05-9ca9-5f1a2dc317db {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.860033] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 873.860033] env[61243]: value = "task-1338993" [ 873.860033] env[61243]: _type = "Task" [ 873.860033] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.871489] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1338993, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.004709] env[61243]: DEBUG nova.network.neutron [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Updated VIF entry in instance network info cache for port 5ea2a618-21af-437b-b35b-69f2a38cc3dc. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 874.005536] env[61243]: DEBUG nova.network.neutron [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Updating instance_info_cache with network_info: [{"id": "5ea2a618-21af-437b-b35b-69f2a38cc3dc", "address": "fa:16:3e:6f:f0:da", "network": {"id": "0788f9ac-5a15-4431-a711-9b4f6cf79fbc", "bridge": null, "label": "tempest-DeleteServersTestJSON-1210957657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "113713170fb94b8a8d7985b9a02faf77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap5ea2a618-21", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.181041] env[61243]: DEBUG oslo_vmware.api [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338982, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.240418] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5424924-7f10-4bcc-ba94-a5542aaab3ff tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 874.240863] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-788b1737-7ddc-4bd9-9978-1520c8021c45 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.248854] env[61243]: DEBUG oslo_vmware.api [None req-c5424924-7f10-4bcc-ba94-a5542aaab3ff tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 874.248854] env[61243]: value = "task-1338994" [ 874.248854] env[61243]: _type = "Task" [ 874.248854] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.257920] env[61243]: DEBUG oslo_vmware.api [None req-c5424924-7f10-4bcc-ba94-a5542aaab3ff tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338994, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.285247] env[61243]: DEBUG oslo_vmware.api [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338991, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.335915] env[61243]: DEBUG nova.compute.utils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 874.337331] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1338992, 'name': ReconfigVM_Task, 'duration_secs': 0.361594} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.338049] env[61243]: DEBUG nova.compute.manager [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 874.338049] env[61243]: DEBUG nova.network.neutron [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 874.339746] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Reconfigured VM instance instance-0000004a to attach disk [datastore2] 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f/8df486c4-d1ce-47b0-81a8-ad9e83eaac7f.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.343041] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65f6accf-328b-4a3d-aaa3-dac4f6661d8d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.349541] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Waiting for the task: (returnval){ [ 874.349541] env[61243]: value = "task-1338995" [ 874.349541] env[61243]: _type = "Task" [ 874.349541] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.353665] env[61243]: DEBUG nova.network.neutron [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 874.361699] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1338995, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.373036] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1338993, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486484} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.375507] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 3f839aa5-2b9a-4807-b63b-931f74455532/3f839aa5-2b9a-4807-b63b-931f74455532.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 874.375771] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 874.376272] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-83f9de4f-dd67-4679-b927-64826a9b8a40 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.383625] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 874.383625] env[61243]: value = "task-1338996" [ 874.383625] env[61243]: _type = "Task" [ 874.383625] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.399428] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1338996, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.419934] env[61243]: DEBUG nova.policy [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7847c54c692c4a8f9ddf93aa32e88690', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4fcef2317fdf47f8ac820aeef1765fef', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 874.483505] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e09ef140-7513-452f-829e-2105aa94a4d9 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.509342] env[61243]: DEBUG oslo_concurrency.lockutils [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] Releasing lock "refresh_cache-ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.509647] env[61243]: DEBUG nova.compute.manager [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Received event network-vif-plugged-c706f966-e43c-4d65-b536-f61daa91c8ef {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.509889] env[61243]: DEBUG oslo_concurrency.lockutils [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] Acquiring lock "783d861c-8930-406e-8986-ef995de4c9dd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.510122] env[61243]: DEBUG oslo_concurrency.lockutils [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] Lock "783d861c-8930-406e-8986-ef995de4c9dd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.510295] env[61243]: DEBUG oslo_concurrency.lockutils [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] Lock "783d861c-8930-406e-8986-ef995de4c9dd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.510472] env[61243]: DEBUG nova.compute.manager [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] No waiting events found dispatching network-vif-plugged-c706f966-e43c-4d65-b536-f61daa91c8ef {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 874.510728] env[61243]: WARNING nova.compute.manager [req-7d833b14-e6ea-423a-9bd2-a247d8ae3869 req-a4be445d-1078-4668-ac6e-ec239edc79a1 service nova] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Received unexpected event network-vif-plugged-c706f966-e43c-4d65-b536-f61daa91c8ef for instance with vm_state building and task_state spawning. [ 874.609841] env[61243]: DEBUG nova.network.neutron [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Updating instance_info_cache with network_info: [{"id": "c706f966-e43c-4d65-b536-f61daa91c8ef", "address": "fa:16:3e:73:a3:9a", "network": {"id": "2737f3f1-a8da-4b5f-a33d-c768f14207e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1004846100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3248cca8a8aa4cad84b8bd5dcf7ed0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc706f966-e4", "ovs_interfaceid": "c706f966-e43c-4d65-b536-f61daa91c8ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.682924] env[61243]: DEBUG oslo_vmware.api [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338982, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.713271] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77ca0977-f767-4d14-a0b0-cc6ab28ca863 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.721640] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff2ccde6-0e41-4e0e-82cd-c8d1d931bf3b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.759130] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f4d2c25-6c9e-488a-a414-294934120234 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.769649] env[61243]: DEBUG oslo_vmware.api [None req-c5424924-7f10-4bcc-ba94-a5542aaab3ff tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338994, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.771033] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da1aab8-65ca-43f4-9c26-bcd8787d010e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.787902] env[61243]: DEBUG nova.compute.provider_tree [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 874.795541] env[61243]: DEBUG oslo_vmware.api [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1338991, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.605976} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.795541] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 874.795541] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 874.795541] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 874.795541] env[61243]: INFO nova.compute.manager [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Took 1.83 seconds to destroy the instance on the hypervisor. [ 874.795541] env[61243]: DEBUG oslo.service.loopingcall [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.795754] env[61243]: DEBUG nova.compute.manager [-] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 874.795903] env[61243]: DEBUG nova.network.neutron [-] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 874.830958] env[61243]: DEBUG nova.network.neutron [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Successfully created port: 15828714-3ad8-4f57-9307-10b1ac64495f {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 874.842274] env[61243]: DEBUG nova.compute.manager [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 874.862668] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1338995, 'name': Rename_Task, 'duration_secs': 0.148684} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.862984] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 874.863309] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d4418a3-8210-437d-af76-b1b8a0fc398b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.872386] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Waiting for the task: (returnval){ [ 874.872386] env[61243]: value = "task-1338997" [ 874.872386] env[61243]: _type = "Task" [ 874.872386] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.879802] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1338997, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.895387] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1338996, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074327} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.896394] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 874.897205] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8141a5-cffc-4829-b136-338879cdfb9b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.923604] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 3f839aa5-2b9a-4807-b63b-931f74455532/3f839aa5-2b9a-4807-b63b-931f74455532.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 874.926494] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93198beb-dde7-476a-bbe0-ee279080936a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.948835] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 874.948835] env[61243]: value = "task-1338998" [ 874.948835] env[61243]: _type = "Task" [ 874.948835] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.957854] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1338998, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.116440] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Releasing lock "refresh_cache-783d861c-8930-406e-8986-ef995de4c9dd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.116587] env[61243]: DEBUG nova.compute.manager [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Instance network_info: |[{"id": "c706f966-e43c-4d65-b536-f61daa91c8ef", "address": "fa:16:3e:73:a3:9a", "network": {"id": "2737f3f1-a8da-4b5f-a33d-c768f14207e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1004846100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3248cca8a8aa4cad84b8bd5dcf7ed0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc706f966-e4", "ovs_interfaceid": "c706f966-e43c-4d65-b536-f61daa91c8ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 875.117022] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:a3:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '26472e27-9835-4f87-ab7f-ca24dfee4e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c706f966-e43c-4d65-b536-f61daa91c8ef', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.125725] env[61243]: DEBUG oslo.service.loopingcall [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.125972] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 875.126240] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-263759c1-8c1d-4427-a044-57da9db97fbb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.147442] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.147442] env[61243]: value = "task-1338999" [ 875.147442] env[61243]: _type = "Task" [ 875.147442] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.158790] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338999, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.184108] env[61243]: DEBUG oslo_vmware.api [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338982, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.272556] env[61243]: DEBUG oslo_vmware.api [None req-c5424924-7f10-4bcc-ba94-a5542aaab3ff tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1338994, 'name': PowerOnVM_Task, 'duration_secs': 0.806166} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.273025] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5424924-7f10-4bcc-ba94-a5542aaab3ff tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 875.273341] env[61243]: DEBUG nova.compute.manager [None req-c5424924-7f10-4bcc-ba94-a5542aaab3ff tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 875.274471] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3691bd6d-ce9f-447f-b45f-7ad5ffb7a3f9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.302329] env[61243]: DEBUG nova.compute.manager [req-ae6b128f-167c-41d8-a56d-3c0ef6cda6c9 req-cf834cbe-8e58-4a9b-90ce-0da4f5568376 service nova] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Received event network-changed-c706f966-e43c-4d65-b536-f61daa91c8ef {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.302329] env[61243]: DEBUG nova.compute.manager [req-ae6b128f-167c-41d8-a56d-3c0ef6cda6c9 req-cf834cbe-8e58-4a9b-90ce-0da4f5568376 service nova] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Refreshing instance network info cache due to event network-changed-c706f966-e43c-4d65-b536-f61daa91c8ef. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 875.302329] env[61243]: DEBUG oslo_concurrency.lockutils [req-ae6b128f-167c-41d8-a56d-3c0ef6cda6c9 req-cf834cbe-8e58-4a9b-90ce-0da4f5568376 service nova] Acquiring lock "refresh_cache-783d861c-8930-406e-8986-ef995de4c9dd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.302689] env[61243]: DEBUG oslo_concurrency.lockutils [req-ae6b128f-167c-41d8-a56d-3c0ef6cda6c9 req-cf834cbe-8e58-4a9b-90ce-0da4f5568376 service nova] Acquired lock "refresh_cache-783d861c-8930-406e-8986-ef995de4c9dd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.302932] env[61243]: DEBUG nova.network.neutron [req-ae6b128f-167c-41d8-a56d-3c0ef6cda6c9 req-cf834cbe-8e58-4a9b-90ce-0da4f5568376 service nova] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Refreshing network info cache for port c706f966-e43c-4d65-b536-f61daa91c8ef {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 875.339698] env[61243]: DEBUG nova.scheduler.client.report [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 94 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 875.339983] env[61243]: DEBUG nova.compute.provider_tree [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 94 to 95 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 875.340186] env[61243]: DEBUG nova.compute.provider_tree [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 875.382040] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1338997, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.459453] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1338998, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.658135] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338999, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.682216] env[61243]: DEBUG oslo_vmware.api [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338982, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.765045] env[61243]: DEBUG nova.network.neutron [-] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.845883] env[61243]: DEBUG oslo_concurrency.lockutils [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.012s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.849493] env[61243]: DEBUG oslo_concurrency.lockutils [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.131s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.849926] env[61243]: DEBUG nova.objects.instance [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lazy-loading 'resources' on Instance uuid 1dc8ad6a-3a00-47c6-8985-481e7f1363cd {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.857018] env[61243]: DEBUG nova.compute.manager [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 875.882023] env[61243]: INFO nova.scheduler.client.report [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Deleted allocations for instance 9a66fd05-1880-423a-8f60-4f11f1c75ab3 [ 875.892706] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1338997, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.895643] env[61243]: DEBUG nova.virt.hardware [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 875.896510] env[61243]: DEBUG nova.virt.hardware [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 875.896510] env[61243]: DEBUG nova.virt.hardware [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.896510] env[61243]: DEBUG nova.virt.hardware [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 875.896510] env[61243]: DEBUG nova.virt.hardware [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.896695] env[61243]: DEBUG nova.virt.hardware [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 875.896784] env[61243]: DEBUG nova.virt.hardware [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 875.897610] env[61243]: DEBUG nova.virt.hardware [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 875.897610] env[61243]: DEBUG nova.virt.hardware [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 875.897610] env[61243]: DEBUG nova.virt.hardware [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 875.897610] env[61243]: DEBUG nova.virt.hardware [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.898432] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0032735f-ac18-4339-9849-69c2c93aaca6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.912794] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb7f725-6aa3-4ac5-8233-d7f9c15081b5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.962778] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1338998, 'name': ReconfigVM_Task, 'duration_secs': 0.983491} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.963097] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 3f839aa5-2b9a-4807-b63b-931f74455532/3f839aa5-2b9a-4807-b63b-931f74455532.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 875.963833] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a32539e6-b3a1-49c2-8f61-9d61371041cb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.970336] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 875.970336] env[61243]: value = "task-1339000" [ 875.970336] env[61243]: _type = "Task" [ 875.970336] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.978749] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339000, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.057786] env[61243]: DEBUG nova.network.neutron [req-ae6b128f-167c-41d8-a56d-3c0ef6cda6c9 req-cf834cbe-8e58-4a9b-90ce-0da4f5568376 service nova] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Updated VIF entry in instance network info cache for port c706f966-e43c-4d65-b536-f61daa91c8ef. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 876.058251] env[61243]: DEBUG nova.network.neutron [req-ae6b128f-167c-41d8-a56d-3c0ef6cda6c9 req-cf834cbe-8e58-4a9b-90ce-0da4f5568376 service nova] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Updating instance_info_cache with network_info: [{"id": "c706f966-e43c-4d65-b536-f61daa91c8ef", "address": "fa:16:3e:73:a3:9a", "network": {"id": "2737f3f1-a8da-4b5f-a33d-c768f14207e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1004846100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3248cca8a8aa4cad84b8bd5dcf7ed0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc706f966-e4", "ovs_interfaceid": "c706f966-e43c-4d65-b536-f61daa91c8ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.160056] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1338999, 'name': CreateVM_Task, 'duration_secs': 0.812392} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.160056] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 876.160056] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.160056] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.160056] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 876.160420] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77be140c-4d41-44eb-b9d4-41abaa01d52a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.164964] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 876.164964] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524670fa-7b64-4b9c-9720-c8369faa6066" [ 876.164964] env[61243]: _type = "Task" [ 876.164964] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.172581] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524670fa-7b64-4b9c-9720-c8369faa6066, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.181530] env[61243]: DEBUG oslo_vmware.api [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338982, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.268020] env[61243]: INFO nova.compute.manager [-] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Took 1.47 seconds to deallocate network for instance. [ 876.385955] env[61243]: DEBUG oslo_vmware.api [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1338997, 'name': PowerOnVM_Task, 'duration_secs': 1.136249} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.385955] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 876.385955] env[61243]: INFO nova.compute.manager [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Took 8.88 seconds to spawn the instance on the hypervisor. [ 876.386160] env[61243]: DEBUG nova.compute.manager [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 876.386905] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05f13eb-637e-429f-968a-eff9adc95024 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.399865] env[61243]: DEBUG oslo_concurrency.lockutils [None req-29778e15-bfe0-485f-9fbe-838d170a3237 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "9a66fd05-1880-423a-8f60-4f11f1c75ab3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.518s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.479839] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339000, 'name': Rename_Task, 'duration_secs': 0.13614} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.481659] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 876.482089] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-610a30c0-4e0d-4c7a-9027-efcda7eb2144 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.488664] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 876.488664] env[61243]: value = "task-1339001" [ 876.488664] env[61243]: _type = "Task" [ 876.488664] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.498713] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339001, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.554169] env[61243]: DEBUG nova.network.neutron [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Successfully updated port: 15828714-3ad8-4f57-9307-10b1ac64495f {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 876.560457] env[61243]: DEBUG oslo_concurrency.lockutils [req-ae6b128f-167c-41d8-a56d-3c0ef6cda6c9 req-cf834cbe-8e58-4a9b-90ce-0da4f5568376 service nova] Releasing lock "refresh_cache-783d861c-8930-406e-8986-ef995de4c9dd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.560754] env[61243]: DEBUG nova.compute.manager [req-ae6b128f-167c-41d8-a56d-3c0ef6cda6c9 req-cf834cbe-8e58-4a9b-90ce-0da4f5568376 service nova] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Received event network-vif-deleted-fe33c43c-f24f-4b9e-9cb9-577bc487106e {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.560999] env[61243]: INFO nova.compute.manager [req-ae6b128f-167c-41d8-a56d-3c0ef6cda6c9 req-cf834cbe-8e58-4a9b-90ce-0da4f5568376 service nova] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Neutron deleted interface fe33c43c-f24f-4b9e-9cb9-577bc487106e; detaching it from the instance and deleting it from the info cache [ 876.561380] env[61243]: DEBUG nova.network.neutron [req-ae6b128f-167c-41d8-a56d-3c0ef6cda6c9 req-cf834cbe-8e58-4a9b-90ce-0da4f5568376 service nova] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.609169] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7091c189-1973-4ff4-9dbc-cc19e89aa436 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.617011] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca30babb-f687-4624-8add-3e7b42108e39 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.647298] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27566493-7013-4afe-ac94-ea825c4a7710 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.654675] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c47f32a-6060-480e-9b40-3c58819bc242 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.667454] env[61243]: DEBUG nova.compute.provider_tree [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 876.676491] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524670fa-7b64-4b9c-9720-c8369faa6066, 'name': SearchDatastore_Task, 'duration_secs': 0.043476} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.680129] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.680390] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.680828] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.680828] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.680940] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.681470] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-17902aa4-b265-4b41-9203-c5eac857b39f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.687621] env[61243]: DEBUG oslo_vmware.api [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338982, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.690481] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.690660] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 876.691355] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90df01a7-463f-43c2-8fc5-03978a8d654d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.696419] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 876.696419] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5274f9cb-cfdf-fba5-ea6b-8a666b33bd9a" [ 876.696419] env[61243]: _type = "Task" [ 876.696419] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.703243] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5274f9cb-cfdf-fba5-ea6b-8a666b33bd9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.774247] env[61243]: DEBUG oslo_concurrency.lockutils [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.909298] env[61243]: INFO nova.compute.manager [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Took 29.48 seconds to build instance. [ 876.999875] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339001, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.056185] env[61243]: DEBUG oslo_concurrency.lockutils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "refresh_cache-6036199d-cda9-42f8-8669-bc099d2b5ffe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.056347] env[61243]: DEBUG oslo_concurrency.lockutils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquired lock "refresh_cache-6036199d-cda9-42f8-8669-bc099d2b5ffe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.056507] env[61243]: DEBUG nova.network.neutron [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 877.063849] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c439eda9-e766-472e-b51f-af0a8f1aa8b2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.073706] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b17a28cf-4b9d-4e83-a868-a4698ae0d988 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.104938] env[61243]: DEBUG nova.compute.manager [req-ae6b128f-167c-41d8-a56d-3c0ef6cda6c9 req-cf834cbe-8e58-4a9b-90ce-0da4f5568376 service nova] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Detach interface failed, port_id=fe33c43c-f24f-4b9e-9cb9-577bc487106e, reason: Instance ca230425-aaa7-4c32-81d3-379cbce0d41b could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 877.187099] env[61243]: DEBUG oslo_vmware.api [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338982, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.206862] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5274f9cb-cfdf-fba5-ea6b-8a666b33bd9a, 'name': SearchDatastore_Task, 'duration_secs': 0.030873} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.207916] env[61243]: DEBUG nova.scheduler.client.report [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 95 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 877.208177] env[61243]: DEBUG nova.compute.provider_tree [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 95 to 96 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 877.208369] env[61243]: DEBUG nova.compute.provider_tree [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 877.213318] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfac3e5b-e8eb-44fb-9e31-98bc3d190ac0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.219295] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 877.219295] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cc8ff7-ab74-6875-9710-177fa496ea8b" [ 877.219295] env[61243]: _type = "Task" [ 877.219295] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.229331] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cc8ff7-ab74-6875-9710-177fa496ea8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.346999] env[61243]: DEBUG nova.compute.manager [req-74df80ac-040d-4dd2-a4e3-80c9f3c870cc req-44e30e1a-c8c5-46e7-b0dd-c5a2feb9236b service nova] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Received event network-vif-plugged-15828714-3ad8-4f57-9307-10b1ac64495f {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.347282] env[61243]: DEBUG oslo_concurrency.lockutils [req-74df80ac-040d-4dd2-a4e3-80c9f3c870cc req-44e30e1a-c8c5-46e7-b0dd-c5a2feb9236b service nova] Acquiring lock "6036199d-cda9-42f8-8669-bc099d2b5ffe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.347544] env[61243]: DEBUG oslo_concurrency.lockutils [req-74df80ac-040d-4dd2-a4e3-80c9f3c870cc req-44e30e1a-c8c5-46e7-b0dd-c5a2feb9236b service nova] Lock "6036199d-cda9-42f8-8669-bc099d2b5ffe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.347764] env[61243]: DEBUG oslo_concurrency.lockutils [req-74df80ac-040d-4dd2-a4e3-80c9f3c870cc req-44e30e1a-c8c5-46e7-b0dd-c5a2feb9236b service nova] Lock "6036199d-cda9-42f8-8669-bc099d2b5ffe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.348015] env[61243]: DEBUG nova.compute.manager [req-74df80ac-040d-4dd2-a4e3-80c9f3c870cc req-44e30e1a-c8c5-46e7-b0dd-c5a2feb9236b service nova] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] No waiting events found dispatching network-vif-plugged-15828714-3ad8-4f57-9307-10b1ac64495f {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 877.348239] env[61243]: WARNING nova.compute.manager [req-74df80ac-040d-4dd2-a4e3-80c9f3c870cc req-44e30e1a-c8c5-46e7-b0dd-c5a2feb9236b service nova] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Received unexpected event network-vif-plugged-15828714-3ad8-4f57-9307-10b1ac64495f for instance with vm_state building and task_state spawning. [ 877.348471] env[61243]: DEBUG nova.compute.manager [req-74df80ac-040d-4dd2-a4e3-80c9f3c870cc req-44e30e1a-c8c5-46e7-b0dd-c5a2feb9236b service nova] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Received event network-changed-15828714-3ad8-4f57-9307-10b1ac64495f {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.348707] env[61243]: DEBUG nova.compute.manager [req-74df80ac-040d-4dd2-a4e3-80c9f3c870cc req-44e30e1a-c8c5-46e7-b0dd-c5a2feb9236b service nova] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Refreshing instance network info cache due to event network-changed-15828714-3ad8-4f57-9307-10b1ac64495f. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 877.348929] env[61243]: DEBUG oslo_concurrency.lockutils [req-74df80ac-040d-4dd2-a4e3-80c9f3c870cc req-44e30e1a-c8c5-46e7-b0dd-c5a2feb9236b service nova] Acquiring lock "refresh_cache-6036199d-cda9-42f8-8669-bc099d2b5ffe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.411214] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7cd467e-0592-498d-9583-89babb215bc4 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Lock "8df486c4-d1ce-47b0-81a8-ad9e83eaac7f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.768s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.499475] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339001, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.593383] env[61243]: DEBUG nova.network.neutron [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 877.688327] env[61243]: DEBUG oslo_vmware.api [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338982, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.714753] env[61243]: DEBUG oslo_concurrency.lockutils [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.865s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.717973] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.987s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.718444] env[61243]: DEBUG nova.objects.instance [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Lazy-loading 'resources' on Instance uuid 6f33b303-1ecd-478e-8fa3-33a5f9a5af50 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.730421] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cc8ff7-ab74-6875-9710-177fa496ea8b, 'name': SearchDatastore_Task, 'duration_secs': 0.010436} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.731691] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.732387] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 783d861c-8930-406e-8986-ef995de4c9dd/783d861c-8930-406e-8986-ef995de4c9dd.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 877.732896] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5303c14-c243-49e4-ba6a-f9ab49068a14 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.737040] env[61243]: DEBUG nova.network.neutron [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Updating instance_info_cache with network_info: [{"id": "15828714-3ad8-4f57-9307-10b1ac64495f", "address": "fa:16:3e:40:09:f1", "network": {"id": "8e2b6835-484b-4346-8bd1-034c53671110", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710127686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4fcef2317fdf47f8ac820aeef1765fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15828714-3a", "ovs_interfaceid": "15828714-3ad8-4f57-9307-10b1ac64495f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.739978] env[61243]: INFO nova.scheduler.client.report [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Deleted allocations for instance 1dc8ad6a-3a00-47c6-8985-481e7f1363cd [ 877.747409] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 877.747409] env[61243]: value = "task-1339002" [ 877.747409] env[61243]: _type = "Task" [ 877.747409] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.760595] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339002, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.005540] env[61243]: DEBUG oslo_vmware.api [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339001, 'name': PowerOnVM_Task, 'duration_secs': 1.479437} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.005958] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 878.006298] env[61243]: INFO nova.compute.manager [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Took 8.09 seconds to spawn the instance on the hypervisor. [ 878.006569] env[61243]: DEBUG nova.compute.manager [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.007631] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c7317f-5956-404e-ac14-25e6af9077b4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.191651] env[61243]: DEBUG oslo_vmware.api [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1338982, 'name': ReconfigVM_Task, 'duration_secs': 5.761896} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.192139] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.192489] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Reconfigured VM to detach interface {{(pid=61243) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 878.243641] env[61243]: DEBUG oslo_concurrency.lockutils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Releasing lock "refresh_cache-6036199d-cda9-42f8-8669-bc099d2b5ffe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.244231] env[61243]: DEBUG nova.compute.manager [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Instance network_info: |[{"id": "15828714-3ad8-4f57-9307-10b1ac64495f", "address": "fa:16:3e:40:09:f1", "network": {"id": "8e2b6835-484b-4346-8bd1-034c53671110", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710127686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4fcef2317fdf47f8ac820aeef1765fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15828714-3a", "ovs_interfaceid": "15828714-3ad8-4f57-9307-10b1ac64495f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 878.244361] env[61243]: DEBUG oslo_concurrency.lockutils [req-74df80ac-040d-4dd2-a4e3-80c9f3c870cc req-44e30e1a-c8c5-46e7-b0dd-c5a2feb9236b service nova] Acquired lock "refresh_cache-6036199d-cda9-42f8-8669-bc099d2b5ffe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.244974] env[61243]: DEBUG nova.network.neutron [req-74df80ac-040d-4dd2-a4e3-80c9f3c870cc req-44e30e1a-c8c5-46e7-b0dd-c5a2feb9236b service nova] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Refreshing network info cache for port 15828714-3ad8-4f57-9307-10b1ac64495f {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 878.249680] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:09:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '415e68b4-3766-4359-afe2-f8563910d98c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '15828714-3ad8-4f57-9307-10b1ac64495f', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.267992] env[61243]: DEBUG oslo.service.loopingcall [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.278831] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 878.279300] env[61243]: DEBUG oslo_concurrency.lockutils [None req-322103c8-7a21-46c3-bfad-b337d30dee18 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "1dc8ad6a-3a00-47c6-8985-481e7f1363cd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.095s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.285582] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-569b5579-454e-462b-81ae-162062ad42ba {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.312973] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339002, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.315396] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 878.315396] env[61243]: value = "task-1339003" [ 878.315396] env[61243]: _type = "Task" [ 878.315396] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.324223] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339003, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.454602] env[61243]: DEBUG nova.compute.manager [req-84713568-764b-47b6-a923-fe9b94845177 req-5875d40b-07c1-44a0-8187-20fd9d5c5208 service nova] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Received event network-changed-5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.455276] env[61243]: DEBUG nova.compute.manager [req-84713568-764b-47b6-a923-fe9b94845177 req-5875d40b-07c1-44a0-8187-20fd9d5c5208 service nova] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Refreshing instance network info cache due to event network-changed-5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 878.455653] env[61243]: DEBUG oslo_concurrency.lockutils [req-84713568-764b-47b6-a923-fe9b94845177 req-5875d40b-07c1-44a0-8187-20fd9d5c5208 service nova] Acquiring lock "refresh_cache-8df486c4-d1ce-47b0-81a8-ad9e83eaac7f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.455922] env[61243]: DEBUG oslo_concurrency.lockutils [req-84713568-764b-47b6-a923-fe9b94845177 req-5875d40b-07c1-44a0-8187-20fd9d5c5208 service nova] Acquired lock "refresh_cache-8df486c4-d1ce-47b0-81a8-ad9e83eaac7f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.456224] env[61243]: DEBUG nova.network.neutron [req-84713568-764b-47b6-a923-fe9b94845177 req-5875d40b-07c1-44a0-8187-20fd9d5c5208 service nova] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Refreshing network info cache for port 5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 878.529147] env[61243]: INFO nova.compute.manager [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Took 29.11 seconds to build instance. [ 878.569748] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94dfa1ad-9c24-4eec-a72f-af2f591cfd6d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.580154] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0a9dcf-efbc-4074-b7a4-e0904624a897 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.612350] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aaf648d-ffb6-4e15-84cc-226b8d174014 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.621717] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5118a524-0991-46b1-9bd2-ed33ad4572eb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.637031] env[61243]: DEBUG nova.compute.provider_tree [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 878.788546] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339002, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.606819} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.788834] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 783d861c-8930-406e-8986-ef995de4c9dd/783d861c-8930-406e-8986-ef995de4c9dd.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 878.789079] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.789449] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-63d7576e-815b-4a45-992c-08cc1a162673 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.796088] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 878.796088] env[61243]: value = "task-1339004" [ 878.796088] env[61243]: _type = "Task" [ 878.796088] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.804870] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339004, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.824142] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339003, 'name': CreateVM_Task, 'duration_secs': 0.451621} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.824326] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 878.825073] env[61243]: DEBUG oslo_concurrency.lockutils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.825256] env[61243]: DEBUG oslo_concurrency.lockutils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.825585] env[61243]: DEBUG oslo_concurrency.lockutils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 878.825848] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8fa50ce-fa6f-4470-9acd-b14e00c1e2ef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.830355] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 878.830355] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522719db-0bc2-6e2e-6fe4-5ae4dc00eb03" [ 878.830355] env[61243]: _type = "Task" [ 878.830355] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.838219] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522719db-0bc2-6e2e-6fe4-5ae4dc00eb03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.012437] env[61243]: DEBUG nova.network.neutron [req-74df80ac-040d-4dd2-a4e3-80c9f3c870cc req-44e30e1a-c8c5-46e7-b0dd-c5a2feb9236b service nova] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Updated VIF entry in instance network info cache for port 15828714-3ad8-4f57-9307-10b1ac64495f. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 879.012819] env[61243]: DEBUG nova.network.neutron [req-74df80ac-040d-4dd2-a4e3-80c9f3c870cc req-44e30e1a-c8c5-46e7-b0dd-c5a2feb9236b service nova] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Updating instance_info_cache with network_info: [{"id": "15828714-3ad8-4f57-9307-10b1ac64495f", "address": "fa:16:3e:40:09:f1", "network": {"id": "8e2b6835-484b-4346-8bd1-034c53671110", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710127686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4fcef2317fdf47f8ac820aeef1765fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15828714-3a", "ovs_interfaceid": "15828714-3ad8-4f57-9307-10b1ac64495f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.030831] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b46dd379-8a52-4634-8cf6-2a14d02d8271 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "3f839aa5-2b9a-4807-b63b-931f74455532" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.954s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.158504] env[61243]: ERROR nova.scheduler.client.report [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] [req-7392eb59-e3c4-4993-9dc4-999c79f94aee] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7392eb59-e3c4-4993-9dc4-999c79f94aee"}]} [ 879.174697] env[61243]: DEBUG nova.scheduler.client.report [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 879.191715] env[61243]: DEBUG nova.scheduler.client.report [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 879.191966] env[61243]: DEBUG nova.compute.provider_tree [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 879.203540] env[61243]: DEBUG nova.scheduler.client.report [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 879.234884] env[61243]: DEBUG nova.scheduler.client.report [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 879.253270] env[61243]: DEBUG nova.network.neutron [req-84713568-764b-47b6-a923-fe9b94845177 req-5875d40b-07c1-44a0-8187-20fd9d5c5208 service nova] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Updated VIF entry in instance network info cache for port 5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 879.253546] env[61243]: DEBUG nova.network.neutron [req-84713568-764b-47b6-a923-fe9b94845177 req-5875d40b-07c1-44a0-8187-20fd9d5c5208 service nova] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Updating instance_info_cache with network_info: [{"id": "5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7", "address": "fa:16:3e:90:9c:10", "network": {"id": "b77f74a5-6460-46f1-910d-b64eaca5e69f", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1175228587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf51131cc0e247669ddf5495ffebd8d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cebc48c-6a77-46bf-9c12-ac130e4d7d76", "external-id": "nsx-vlan-transportzone-382", "segmentation_id": 382, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ae039d5-ce", "ovs_interfaceid": "5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.306729] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339004, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085928} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.308116] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.308116] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba6d8926-db56-4a77-8fc9-fe3ca9540f37 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.334091] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 783d861c-8930-406e-8986-ef995de4c9dd/783d861c-8930-406e-8986-ef995de4c9dd.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.335985] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3867ab83-8085-4d67-b090-10c518f92b3d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.361679] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522719db-0bc2-6e2e-6fe4-5ae4dc00eb03, 'name': SearchDatastore_Task, 'duration_secs': 0.010033} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.363337] env[61243]: DEBUG oslo_concurrency.lockutils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.363337] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.363629] env[61243]: DEBUG oslo_concurrency.lockutils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.364600] env[61243]: DEBUG oslo_concurrency.lockutils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.364600] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.364600] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 879.364600] env[61243]: value = "task-1339005" [ 879.364600] env[61243]: _type = "Task" [ 879.364600] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.364600] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-377f1447-c77e-4cb2-9fee-80f7d1561d2c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.380516] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339005, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.387670] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.388042] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 879.392020] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5ee6ff8-2022-4722-a3db-bd18bdf7ee9f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.401156] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 879.401156] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525019d5-5ff0-a828-4a64-e7a8ac2a6167" [ 879.401156] env[61243]: _type = "Task" [ 879.401156] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.409550] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525019d5-5ff0-a828-4a64-e7a8ac2a6167, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.446492] env[61243]: DEBUG nova.compute.manager [req-b5bd9e3c-58d6-498e-8004-b30ce74855dd req-b49f5c57-d167-40ba-b575-67cd21613496 service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Received event network-vif-deleted-7138c7dd-4984-4d4f-a631-0d0f531c0044 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.446492] env[61243]: INFO nova.compute.manager [req-b5bd9e3c-58d6-498e-8004-b30ce74855dd req-b49f5c57-d167-40ba-b575-67cd21613496 service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Neutron deleted interface 7138c7dd-4984-4d4f-a631-0d0f531c0044; detaching it from the instance and deleting it from the info cache [ 879.446492] env[61243]: DEBUG nova.network.neutron [req-b5bd9e3c-58d6-498e-8004-b30ce74855dd req-b49f5c57-d167-40ba-b575-67cd21613496 service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Updating instance_info_cache with network_info: [{"id": "4cd8c409-8daa-4f53-ac5f-d2ac25de247b", "address": "fa:16:3e:11:60:2d", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4cd8c409-8d", "ovs_interfaceid": "4cd8c409-8daa-4f53-ac5f-d2ac25de247b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.498149] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.498499] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.498634] env[61243]: DEBUG nova.network.neutron [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 879.516895] env[61243]: DEBUG oslo_concurrency.lockutils [req-74df80ac-040d-4dd2-a4e3-80c9f3c870cc req-44e30e1a-c8c5-46e7-b0dd-c5a2feb9236b service nova] Releasing lock "refresh_cache-6036199d-cda9-42f8-8669-bc099d2b5ffe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.603028] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "f8816b86-10c8-4d64-9659-fbc301893d46" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.603361] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "f8816b86-10c8-4d64-9659-fbc301893d46" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.622918] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3bd64e2-7400-424d-a9f7-0d8746bf9b24 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.633580] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5275ba9-41a0-43b6-8d87-54df25a36879 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.641209] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "d95a4d61-2bff-4d89-9ab8-f71e558f3409" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.641474] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "d95a4d61-2bff-4d89-9ab8-f71e558f3409" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.675382] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd619c7-03d8-4284-89d6-11334729cde0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.683554] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8733fd20-9b8e-40f3-ba51-f7126e26f4b0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.699435] env[61243]: DEBUG nova.compute.provider_tree [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 879.759034] env[61243]: DEBUG oslo_concurrency.lockutils [req-84713568-764b-47b6-a923-fe9b94845177 req-5875d40b-07c1-44a0-8187-20fd9d5c5208 service nova] Releasing lock "refresh_cache-8df486c4-d1ce-47b0-81a8-ad9e83eaac7f" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.876972] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339005, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.911330] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525019d5-5ff0-a828-4a64-e7a8ac2a6167, 'name': SearchDatastore_Task, 'duration_secs': 0.025657} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.912126] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce952e9d-735b-46e2-8f09-abe753e18dc1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.917095] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 879.917095] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52dc25f7-23f6-a023-9f3a-28826a4b29df" [ 879.917095] env[61243]: _type = "Task" [ 879.917095] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.924079] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52dc25f7-23f6-a023-9f3a-28826a4b29df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.949521] env[61243]: DEBUG oslo_concurrency.lockutils [req-b5bd9e3c-58d6-498e-8004-b30ce74855dd req-b49f5c57-d167-40ba-b575-67cd21613496 service nova] Acquiring lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.949828] env[61243]: DEBUG oslo_concurrency.lockutils [req-b5bd9e3c-58d6-498e-8004-b30ce74855dd req-b49f5c57-d167-40ba-b575-67cd21613496 service nova] Acquired lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.950736] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cedfad88-86c3-44f1-8b41-9af354526ee6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.967041] env[61243]: DEBUG oslo_concurrency.lockutils [req-b5bd9e3c-58d6-498e-8004-b30ce74855dd req-b49f5c57-d167-40ba-b575-67cd21613496 service nova] Releasing lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.967364] env[61243]: WARNING nova.compute.manager [req-b5bd9e3c-58d6-498e-8004-b30ce74855dd req-b49f5c57-d167-40ba-b575-67cd21613496 service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Detach interface failed, port_id=7138c7dd-4984-4d4f-a631-0d0f531c0044, reason: No device with interface-id 7138c7dd-4984-4d4f-a631-0d0f531c0044 exists on VM: nova.exception.NotFound: No device with interface-id 7138c7dd-4984-4d4f-a631-0d0f531c0044 exists on VM [ 880.106600] env[61243]: DEBUG nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 880.146022] env[61243]: DEBUG nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 880.148707] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.148950] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.149175] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.149362] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.149668] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.151934] env[61243]: INFO nova.compute.manager [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Terminating instance [ 880.153727] env[61243]: DEBUG nova.compute.manager [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 880.153921] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 880.154734] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f99a92a-3252-4ca7-baef-a86c0ea45a2b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.165483] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 880.165483] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0607ae71-3bb1-4d87-a3f1-4e1036882d39 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.172060] env[61243]: DEBUG oslo_vmware.api [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 880.172060] env[61243]: value = "task-1339006" [ 880.172060] env[61243]: _type = "Task" [ 880.172060] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.179599] env[61243]: DEBUG oslo_vmware.api [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339006, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.231620] env[61243]: DEBUG nova.scheduler.client.report [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 98 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 880.231904] env[61243]: DEBUG nova.compute.provider_tree [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 98 to 99 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 880.232101] env[61243]: DEBUG nova.compute.provider_tree [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 880.235974] env[61243]: DEBUG nova.network.neutron [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Updating instance_info_cache with network_info: [{"id": "4cd8c409-8daa-4f53-ac5f-d2ac25de247b", "address": "fa:16:3e:11:60:2d", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4cd8c409-8d", "ovs_interfaceid": "4cd8c409-8daa-4f53-ac5f-d2ac25de247b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.379352] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339005, 'name': ReconfigVM_Task, 'duration_secs': 0.774765} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.380462] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 783d861c-8930-406e-8986-ef995de4c9dd/783d861c-8930-406e-8986-ef995de4c9dd.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.380462] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ecc6b0ba-aa9c-4e30-99e2-c09109ca854a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.388086] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 880.388086] env[61243]: value = "task-1339007" [ 880.388086] env[61243]: _type = "Task" [ 880.388086] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.395932] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339007, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.428023] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52dc25f7-23f6-a023-9f3a-28826a4b29df, 'name': SearchDatastore_Task, 'duration_secs': 0.010218} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.428023] env[61243]: DEBUG oslo_concurrency.lockutils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.428232] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 6036199d-cda9-42f8-8669-bc099d2b5ffe/6036199d-cda9-42f8-8669-bc099d2b5ffe.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 880.428520] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b7b4ac56-549f-46ba-9439-7a9c39caff99 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.435469] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 880.435469] env[61243]: value = "task-1339008" [ 880.435469] env[61243]: _type = "Task" [ 880.435469] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.443260] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339008, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.560317] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.560755] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.628861] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.669256] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.681840] env[61243]: DEBUG oslo_vmware.api [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339006, 'name': PowerOffVM_Task, 'duration_secs': 0.343283} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.682863] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 880.682863] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 880.682863] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1762f975-8e0d-4373-82b7-4c0aa38f5a24 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.738533] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.021s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.741036] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "refresh_cache-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.744780] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.981s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.744780] env[61243]: DEBUG nova.objects.instance [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lazy-loading 'resources' on Instance uuid 96315d9e-4eda-4e3a-af0d-bdc52ab181e3 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.761930] env[61243]: INFO nova.scheduler.client.report [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Deleted allocations for instance 6f33b303-1ecd-478e-8fa3-33a5f9a5af50 [ 880.807088] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 880.807088] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 880.807481] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Deleting the datastore file [datastore2] 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 880.807642] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-53c28f55-518e-4272-a952-c5f30e6a3d25 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.814761] env[61243]: DEBUG oslo_vmware.api [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 880.814761] env[61243]: value = "task-1339010" [ 880.814761] env[61243]: _type = "Task" [ 880.814761] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.826074] env[61243]: DEBUG oslo_vmware.api [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339010, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.898874] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339007, 'name': Rename_Task, 'duration_secs': 0.157322} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.899351] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 880.899785] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e3d4a963-a9ee-451a-a432-5e8bc1ea78b0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.908392] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 880.908392] env[61243]: value = "task-1339011" [ 880.908392] env[61243]: _type = "Task" [ 880.908392] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.918233] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339011, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.945825] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339008, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.064021] env[61243]: DEBUG nova.compute.utils [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 881.246339] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c6324e60-f61d-4543-b386-81dba84f4b1c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00-7138c7dd-4984-4d4f-a631-0d0f531c0044" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.671s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.273087] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c968146-c53d-44e9-bb5a-b8517b29ff69 tempest-AttachInterfacesV270Test-1058625807 tempest-AttachInterfacesV270Test-1058625807-project-member] Lock "6f33b303-1ecd-478e-8fa3-33a5f9a5af50" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.010s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.326217] env[61243]: DEBUG oslo_vmware.api [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339010, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.420587] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339011, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.447428] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339008, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516711} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.447625] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 6036199d-cda9-42f8-8669-bc099d2b5ffe/6036199d-cda9-42f8-8669-bc099d2b5ffe.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 881.447899] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.448193] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-450852f8-c046-416d-8825-aa48715fccfe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.456883] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 881.456883] env[61243]: value = "task-1339012" [ 881.456883] env[61243]: _type = "Task" [ 881.456883] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.465230] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339012, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.491369] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c332c314-ca97-4e58-ba74-39afda0c3370 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.498889] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9560984-fb01-4df6-91c5-d6018266044a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.535499] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dac10c4-05cd-4b77-b595-2cdf5784c50e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.539900] env[61243]: DEBUG nova.compute.manager [req-3dd60698-25c7-498d-a51e-8527be4c95ce req-00d92d1b-9dd1-41d4-9b15-7074de99aac3 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Received event network-changed-a0b324d6-6a73-4956-85c3-38f755003b74 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.540180] env[61243]: DEBUG nova.compute.manager [req-3dd60698-25c7-498d-a51e-8527be4c95ce req-00d92d1b-9dd1-41d4-9b15-7074de99aac3 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Refreshing instance network info cache due to event network-changed-a0b324d6-6a73-4956-85c3-38f755003b74. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 881.541044] env[61243]: DEBUG oslo_concurrency.lockutils [req-3dd60698-25c7-498d-a51e-8527be4c95ce req-00d92d1b-9dd1-41d4-9b15-7074de99aac3 service nova] Acquiring lock "refresh_cache-3f839aa5-2b9a-4807-b63b-931f74455532" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.541044] env[61243]: DEBUG oslo_concurrency.lockutils [req-3dd60698-25c7-498d-a51e-8527be4c95ce req-00d92d1b-9dd1-41d4-9b15-7074de99aac3 service nova] Acquired lock "refresh_cache-3f839aa5-2b9a-4807-b63b-931f74455532" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.541044] env[61243]: DEBUG nova.network.neutron [req-3dd60698-25c7-498d-a51e-8527be4c95ce req-00d92d1b-9dd1-41d4-9b15-7074de99aac3 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Refreshing network info cache for port a0b324d6-6a73-4956-85c3-38f755003b74 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 881.552304] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05649de7-87fe-4cc1-a3db-ec3cb9c838dd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.567106] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.567907] env[61243]: DEBUG nova.compute.provider_tree [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.824729] env[61243]: DEBUG oslo_vmware.api [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339010, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.660319} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.824995] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 881.825203] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 881.825387] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 881.825590] env[61243]: INFO nova.compute.manager [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Took 1.67 seconds to destroy the instance on the hypervisor. [ 881.825851] env[61243]: DEBUG oslo.service.loopingcall [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.826060] env[61243]: DEBUG nova.compute.manager [-] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 881.826159] env[61243]: DEBUG nova.network.neutron [-] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 881.918519] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339011, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.968203] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339012, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.27944} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.968203] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 881.968408] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db6b0f6e-3945-4aef-afdd-281712eb2137 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.989932] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 6036199d-cda9-42f8-8669-bc099d2b5ffe/6036199d-cda9-42f8-8669-bc099d2b5ffe.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.990262] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af158970-2f4e-44cb-95f0-36183243927c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.010354] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 882.010354] env[61243]: value = "task-1339013" [ 882.010354] env[61243]: _type = "Task" [ 882.010354] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.020120] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339013, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.071910] env[61243]: DEBUG nova.scheduler.client.report [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.419432] env[61243]: DEBUG oslo_vmware.api [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339011, 'name': PowerOnVM_Task, 'duration_secs': 1.234346} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.419721] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 882.419935] env[61243]: INFO nova.compute.manager [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Took 10.07 seconds to spawn the instance on the hypervisor. [ 882.420165] env[61243]: DEBUG nova.compute.manager [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.420963] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbb2b4f9-5a5e-40c5-8501-f8f42b2a4693 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.524265] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339013, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.577057] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.833s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.579274] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 19.340s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.579468] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.579695] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61243) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 882.580078] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.066s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.580194] env[61243]: DEBUG nova.objects.instance [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Lazy-loading 'resources' on Instance uuid f73a4d39-5478-4135-9be6-e59f3e29788d {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 882.584130] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2162952e-9bcf-43ba-89b2-3e6775482179 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.591869] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a26162-1903-43e3-a630-f77d6a3135b4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.612667] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a22a15b-49ff-4d05-a9c3-efcb9a8d7aea {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.616156] env[61243]: INFO nova.scheduler.client.report [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Deleted allocations for instance 96315d9e-4eda-4e3a-af0d-bdc52ab181e3 [ 882.626682] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d8b3c4-43ed-443d-ac22-f654ae36bc57 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.661184] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180448MB free_disk=153GB free_vcpus=48 pci_devices=None {{(pid=61243) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 882.661352] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.673801] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.674063] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.674300] env[61243]: INFO nova.compute.manager [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Attaching volume de767a49-762d-43e5-bf2b-b9f14e719ead to /dev/sdb [ 882.720289] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-884d8202-c432-4063-b220-7030ac0ec015 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.732501] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-befc7c3e-9c43-48d8-badf-d5d876c54885 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.749666] env[61243]: DEBUG nova.virt.block_device [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Updating existing volume attachment record: 49f89fe5-ff04-498a-a708-3e0b80a02d9e {{(pid=61243) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 882.759882] env[61243]: DEBUG nova.network.neutron [req-3dd60698-25c7-498d-a51e-8527be4c95ce req-00d92d1b-9dd1-41d4-9b15-7074de99aac3 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Updated VIF entry in instance network info cache for port a0b324d6-6a73-4956-85c3-38f755003b74. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 882.759882] env[61243]: DEBUG nova.network.neutron [req-3dd60698-25c7-498d-a51e-8527be4c95ce req-00d92d1b-9dd1-41d4-9b15-7074de99aac3 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Updating instance_info_cache with network_info: [{"id": "a0b324d6-6a73-4956-85c3-38f755003b74", "address": "fa:16:3e:16:b3:46", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0b324d6-6a", "ovs_interfaceid": "a0b324d6-6a73-4956-85c3-38f755003b74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.942433] env[61243]: INFO nova.compute.manager [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Took 32.28 seconds to build instance. [ 883.028466] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339013, 'name': ReconfigVM_Task, 'duration_secs': 0.885903} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.028466] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 6036199d-cda9-42f8-8669-bc099d2b5ffe/6036199d-cda9-42f8-8669-bc099d2b5ffe.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.029519] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-519e8c11-b271-4232-a9df-bc24043bc5a4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.037923] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 883.037923] env[61243]: value = "task-1339016" [ 883.037923] env[61243]: _type = "Task" [ 883.037923] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.050250] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339016, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.126971] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4fb8ea1c-2a2c-4e38-afb5-ea52d20e8263 tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "96315d9e-4eda-4e3a-af0d-bdc52ab181e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.167s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.262645] env[61243]: DEBUG oslo_concurrency.lockutils [req-3dd60698-25c7-498d-a51e-8527be4c95ce req-00d92d1b-9dd1-41d4-9b15-7074de99aac3 service nova] Releasing lock "refresh_cache-3f839aa5-2b9a-4807-b63b-931f74455532" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.373704] env[61243]: DEBUG nova.network.neutron [-] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.389549] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1147e72a-3a3e-47fd-9d5f-5d0e48edab78 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.398841] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd6fd79-25a1-4a6f-afc4-91ab59dfed08 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.430026] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97497527-e606-4a1a-95f8-a7314acf7596 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.437554] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9eef38b-ee96-42c9-8f03-7233f09f909f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.450667] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c32d0ba6-2abd-4a5f-84f2-b308572e358e tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "783d861c-8930-406e-8986-ef995de4c9dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.778s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.451626] env[61243]: DEBUG nova.compute.provider_tree [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.549584] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339016, 'name': Rename_Task, 'duration_secs': 0.15499} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.549750] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 883.550045] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5148cec5-21ba-4da4-92ee-68052dc03410 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.556178] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 883.556178] env[61243]: value = "task-1339018" [ 883.556178] env[61243]: _type = "Task" [ 883.556178] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.563809] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339018, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.565833] env[61243]: DEBUG nova.compute.manager [req-f0ae8c37-454b-42d8-b9b8-9752a2223658 req-458014d8-1d1d-4633-8f28-ecbb9d75edfa service nova] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Received event network-vif-deleted-4cd8c409-8daa-4f53-ac5f-d2ac25de247b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.879049] env[61243]: INFO nova.compute.manager [-] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Took 2.05 seconds to deallocate network for instance. [ 883.954458] env[61243]: DEBUG nova.scheduler.client.report [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.066535] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339018, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.204190] env[61243]: DEBUG oslo_concurrency.lockutils [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "729d9bd9-1ab0-47bd-9b7a-11412d3dc608" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.204472] env[61243]: DEBUG oslo_concurrency.lockutils [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "729d9bd9-1ab0-47bd-9b7a-11412d3dc608" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.204694] env[61243]: DEBUG oslo_concurrency.lockutils [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "729d9bd9-1ab0-47bd-9b7a-11412d3dc608-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.204891] env[61243]: DEBUG oslo_concurrency.lockutils [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "729d9bd9-1ab0-47bd-9b7a-11412d3dc608-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.205089] env[61243]: DEBUG oslo_concurrency.lockutils [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "729d9bd9-1ab0-47bd-9b7a-11412d3dc608-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.207129] env[61243]: INFO nova.compute.manager [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Terminating instance [ 884.209307] env[61243]: DEBUG oslo_concurrency.lockutils [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "refresh_cache-729d9bd9-1ab0-47bd-9b7a-11412d3dc608" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.210172] env[61243]: DEBUG oslo_concurrency.lockutils [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquired lock "refresh_cache-729d9bd9-1ab0-47bd-9b7a-11412d3dc608" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.210446] env[61243]: DEBUG nova.network.neutron [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 884.325501] env[61243]: DEBUG nova.compute.manager [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.326428] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-befb61a2-8a46-4a0d-9070-501bdfb3cb0c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.384740] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.459835] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.879s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.461911] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.636s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.462177] env[61243]: DEBUG nova.objects.instance [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lazy-loading 'resources' on Instance uuid 5daccb48-b9bf-423a-ab1a-5aa26d8bec87 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.493072] env[61243]: INFO nova.scheduler.client.report [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Deleted allocations for instance f73a4d39-5478-4135-9be6-e59f3e29788d [ 884.566787] env[61243]: DEBUG oslo_vmware.api [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339018, 'name': PowerOnVM_Task, 'duration_secs': 0.880685} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.567578] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 884.567863] env[61243]: INFO nova.compute.manager [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Took 8.71 seconds to spawn the instance on the hypervisor. [ 884.568084] env[61243]: DEBUG nova.compute.manager [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.568895] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ebad47-5ba6-4b79-a9ba-7e265ed074e9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.736201] env[61243]: DEBUG nova.network.neutron [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.818979] env[61243]: DEBUG nova.network.neutron [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.836970] env[61243]: INFO nova.compute.manager [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] instance snapshotting [ 884.839701] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4028692-96bd-43a1-87c8-ca5942cb36ce {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.860605] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ed4de7-9176-4cec-908d-9933c5341c15 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.000286] env[61243]: DEBUG oslo_concurrency.lockutils [None req-dc8b0494-b45b-4a45-8f0b-9a09adea58e5 tempest-ServersNegativeTestMultiTenantJSON-1348030678 tempest-ServersNegativeTestMultiTenantJSON-1348030678-project-member] Lock "f73a4d39-5478-4135-9be6-e59f3e29788d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.365s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.087984] env[61243]: INFO nova.compute.manager [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Took 31.42 seconds to build instance. [ 885.251233] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d38907f0-3211-43b2-b29c-86ef2dfea612 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.259702] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c901d8-6f87-42e1-b00d-d07d3a8aea29 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.292123] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d5822f-321d-4376-b23e-ff41b3f21e9b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.299191] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77a832f6-abb9-499a-96af-b9f6918a8468 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.315400] env[61243]: DEBUG nova.compute.provider_tree [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.323182] env[61243]: DEBUG oslo_concurrency.lockutils [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Releasing lock "refresh_cache-729d9bd9-1ab0-47bd-9b7a-11412d3dc608" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.323182] env[61243]: DEBUG nova.compute.manager [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 885.323182] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 885.323750] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7c3d83-094c-4fb7-9c82-2d54a0bd32f6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.331997] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 885.332359] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3181dcac-c57a-4dce-9a1f-b54b119b6473 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.340758] env[61243]: DEBUG oslo_vmware.api [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 885.340758] env[61243]: value = "task-1339020" [ 885.340758] env[61243]: _type = "Task" [ 885.340758] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.351588] env[61243]: DEBUG oslo_vmware.api [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1339020, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.372326] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Creating Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 885.372754] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-60c6474c-708d-4cff-a08e-a3f619c6d147 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.385137] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 885.385137] env[61243]: value = "task-1339021" [ 885.385137] env[61243]: _type = "Task" [ 885.385137] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.414684] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339021, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.590879] env[61243]: DEBUG oslo_concurrency.lockutils [None req-49ff67ae-b284-4f1f-b66e-22224da45dff tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "6036199d-cda9-42f8-8669-bc099d2b5ffe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.534s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.818945] env[61243]: DEBUG nova.scheduler.client.report [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 885.855152] env[61243]: DEBUG oslo_vmware.api [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1339020, 'name': PowerOffVM_Task, 'duration_secs': 0.234114} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.855427] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 885.855602] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 885.856015] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4e67bac-5539-477f-b977-c9497e476f3d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.880460] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 885.880720] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 885.880905] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Deleting the datastore file [datastore2] 729d9bd9-1ab0-47bd-9b7a-11412d3dc608 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 885.881201] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-966208e3-a67e-4388-9e16-e41af6dc5bdc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.893482] env[61243]: DEBUG oslo_vmware.api [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for the task: (returnval){ [ 885.893482] env[61243]: value = "task-1339023" [ 885.893482] env[61243]: _type = "Task" [ 885.893482] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.897034] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339021, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.905599] env[61243]: DEBUG oslo_vmware.api [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1339023, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.066486] env[61243]: INFO nova.compute.manager [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Rescuing [ 886.066745] env[61243]: DEBUG oslo_concurrency.lockutils [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "refresh_cache-6036199d-cda9-42f8-8669-bc099d2b5ffe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.066894] env[61243]: DEBUG oslo_concurrency.lockutils [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquired lock "refresh_cache-6036199d-cda9-42f8-8669-bc099d2b5ffe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.067081] env[61243]: DEBUG nova.network.neutron [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 886.325941] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.864s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.329698] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.016s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.329762] env[61243]: DEBUG nova.objects.instance [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lazy-loading 'resources' on Instance uuid ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.351525] env[61243]: INFO nova.scheduler.client.report [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Deleted allocations for instance 5daccb48-b9bf-423a-ab1a-5aa26d8bec87 [ 886.402569] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339021, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.410110] env[61243]: DEBUG oslo_vmware.api [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Task: {'id': task-1339023, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156014} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.410263] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 886.410496] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 886.410727] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 886.410928] env[61243]: INFO nova.compute.manager [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Took 1.09 seconds to destroy the instance on the hypervisor. [ 886.411192] env[61243]: DEBUG oslo.service.loopingcall [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.411412] env[61243]: DEBUG nova.compute.manager [-] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 886.411516] env[61243]: DEBUG nova.network.neutron [-] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 886.432095] env[61243]: DEBUG nova.network.neutron [-] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 886.834583] env[61243]: DEBUG nova.objects.instance [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lazy-loading 'numa_topology' on Instance uuid ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.837456] env[61243]: DEBUG nova.network.neutron [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Updating instance_info_cache with network_info: [{"id": "15828714-3ad8-4f57-9307-10b1ac64495f", "address": "fa:16:3e:40:09:f1", "network": {"id": "8e2b6835-484b-4346-8bd1-034c53671110", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710127686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4fcef2317fdf47f8ac820aeef1765fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15828714-3a", "ovs_interfaceid": "15828714-3ad8-4f57-9307-10b1ac64495f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.860451] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f74158a-f433-4cbb-b780-fe5d22fd0554 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "5daccb48-b9bf-423a-ab1a-5aa26d8bec87" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.304s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.897530] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339021, 'name': CreateSnapshot_Task, 'duration_secs': 1.045635} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.897831] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Created Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 886.898626] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5c8c0c-9f14-4d4a-bc83-39a92e60292f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.938906] env[61243]: DEBUG nova.network.neutron [-] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.294220] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "e9a1c8af-7129-4366-bda1-1c3cf40eeab1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.295747] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "e9a1c8af-7129-4366-bda1-1c3cf40eeab1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.308016] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Volume attach. Driver type: vmdk {{(pid=61243) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 887.308016] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285724', 'volume_id': 'de767a49-762d-43e5-bf2b-b9f14e719ead', 'name': 'volume-de767a49-762d-43e5-bf2b-b9f14e719ead', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '99ce6ca3-a478-4ebe-bf1b-42459aacd9ba', 'attached_at': '', 'detached_at': '', 'volume_id': 'de767a49-762d-43e5-bf2b-b9f14e719ead', 'serial': 'de767a49-762d-43e5-bf2b-b9f14e719ead'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 887.309380] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3bd05fe-a578-4de0-aef1-dcfec67e335e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.337912] env[61243]: DEBUG nova.objects.base [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 887.342597] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf2636c-a73e-4017-a810-54be811ba8aa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.346771] env[61243]: DEBUG oslo_concurrency.lockutils [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Releasing lock "refresh_cache-6036199d-cda9-42f8-8669-bc099d2b5ffe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.385200] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] volume-de767a49-762d-43e5-bf2b-b9f14e719ead/volume-de767a49-762d-43e5-bf2b-b9f14e719ead.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 887.390196] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0031869c-1a2b-42df-bac9-fa99cd9e5197 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.411858] env[61243]: DEBUG oslo_vmware.api [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 887.411858] env[61243]: value = "task-1339024" [ 887.411858] env[61243]: _type = "Task" [ 887.411858] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.422045] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Creating linked-clone VM from snapshot {{(pid=61243) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 887.429975] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-dc6d3841-4c8c-4f0a-bc49-1ac66c300917 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.441366] env[61243]: DEBUG oslo_vmware.api [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339024, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.443197] env[61243]: INFO nova.compute.manager [-] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Took 1.03 seconds to deallocate network for instance. [ 887.445161] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 887.445161] env[61243]: value = "task-1339025" [ 887.445161] env[61243]: _type = "Task" [ 887.445161] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.466393] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339025, 'name': CloneVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.701249] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0b0f64-6b8e-4acd-ba4e-3d6d5a5421d1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.709485] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-800c3bf1-bbeb-4eac-ba70-7b9715d4a678 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.744259] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4baaba4e-3393-40ae-897f-c7384e1308d6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.753270] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e260d05-cb33-4516-adef-3e1e69c6906e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.768009] env[61243]: DEBUG nova.compute.provider_tree [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.797717] env[61243]: DEBUG nova.compute.manager [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 887.929319] env[61243]: DEBUG oslo_vmware.api [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339024, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.929848] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.930119] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7c4926d-8fa6-49ff-8c79-a59c232e7073 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.937794] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 887.937794] env[61243]: value = "task-1339026" [ 887.937794] env[61243]: _type = "Task" [ 887.937794] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.948417] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339026, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.959008] env[61243]: DEBUG oslo_concurrency.lockutils [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.959360] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339025, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.271156] env[61243]: DEBUG nova.scheduler.client.report [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.325255] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.426787] env[61243]: DEBUG oslo_vmware.api [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339024, 'name': ReconfigVM_Task, 'duration_secs': 0.786424} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.427108] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Reconfigured VM instance instance-0000003d to attach disk [datastore2] volume-de767a49-762d-43e5-bf2b-b9f14e719ead/volume-de767a49-762d-43e5-bf2b-b9f14e719ead.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.432633] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51eed160-a14b-457c-8052-7c10056e689e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.456175] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339026, 'name': PowerOffVM_Task, 'duration_secs': 0.209914} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.456485] env[61243]: DEBUG oslo_vmware.api [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 888.456485] env[61243]: value = "task-1339027" [ 888.456485] env[61243]: _type = "Task" [ 888.456485] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.457139] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 888.458023] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89455f6d-a58a-4bc5-8009-097a1ec45a33 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.468152] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339025, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.474310] env[61243]: DEBUG oslo_vmware.api [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339027, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.492604] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db861b2c-0930-4fc9-9d02-a1edc27c9132 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.535101] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 888.535416] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f140d955-8930-4399-9f95-d7eef1b50e13 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.544148] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 888.544148] env[61243]: value = "task-1339028" [ 888.544148] env[61243]: _type = "Task" [ 888.544148] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.555037] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] VM already powered off {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 888.555037] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.555271] env[61243]: DEBUG oslo_concurrency.lockutils [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.555271] env[61243]: DEBUG oslo_concurrency.lockutils [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.555420] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.555688] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bbe13b36-0af4-4260-914b-a4e53bbf4ab9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.597425] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.597632] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 888.598541] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2338a3a5-95bb-4ce6-990e-7f829f20c665 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.606490] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 888.606490] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5229fa02-7c90-3dcf-9c15-45e105918295" [ 888.606490] env[61243]: _type = "Task" [ 888.606490] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.616536] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5229fa02-7c90-3dcf-9c15-45e105918295, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.783255] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.453s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.786083] env[61243]: DEBUG oslo_concurrency.lockutils [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.012s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.786083] env[61243]: DEBUG nova.objects.instance [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lazy-loading 'resources' on Instance uuid ca230425-aaa7-4c32-81d3-379cbce0d41b {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.959765] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339025, 'name': CloneVM_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.970197] env[61243]: DEBUG oslo_vmware.api [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339027, 'name': ReconfigVM_Task, 'duration_secs': 0.17806} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.970530] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285724', 'volume_id': 'de767a49-762d-43e5-bf2b-b9f14e719ead', 'name': 'volume-de767a49-762d-43e5-bf2b-b9f14e719ead', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '99ce6ca3-a478-4ebe-bf1b-42459aacd9ba', 'attached_at': '', 'detached_at': '', 'volume_id': 'de767a49-762d-43e5-bf2b-b9f14e719ead', 'serial': 'de767a49-762d-43e5-bf2b-b9f14e719ead'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 889.118210] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5229fa02-7c90-3dcf-9c15-45e105918295, 'name': SearchDatastore_Task, 'duration_secs': 0.022944} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.119021] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47235090-f6e6-4940-822e-71c70fb26870 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.124903] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 889.124903] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527c7ae1-fa9f-7391-16a2-581878bb20b1" [ 889.124903] env[61243]: _type = "Task" [ 889.124903] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.136887] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527c7ae1-fa9f-7391-16a2-581878bb20b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.298226] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5c71b6e5-3811-4e72-a389-499f9813d669 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 36.273s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.299540] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e09ef140-7513-452f-829e-2105aa94a4d9 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 14.816s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.299777] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e09ef140-7513-452f-829e-2105aa94a4d9 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.299986] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e09ef140-7513-452f-829e-2105aa94a4d9 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.300202] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e09ef140-7513-452f-829e-2105aa94a4d9 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.302876] env[61243]: INFO nova.compute.manager [None req-e09ef140-7513-452f-829e-2105aa94a4d9 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Terminating instance [ 889.305398] env[61243]: DEBUG nova.compute.manager [None req-e09ef140-7513-452f-829e-2105aa94a4d9 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 889.305398] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e09ef140-7513-452f-829e-2105aa94a4d9 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 889.305797] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dea0d40e-4afa-4a68-8717-53e338f8d190 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.317295] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e20b89c-2517-4df4-87f2-78acfd68bc80 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.360515] env[61243]: WARNING nova.virt.vmwareapi.vmops [None req-e09ef140-7513-452f-829e-2105aa94a4d9 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5 could not be found. [ 889.360757] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e09ef140-7513-452f-829e-2105aa94a4d9 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 889.360983] env[61243]: INFO nova.compute.manager [None req-e09ef140-7513-452f-829e-2105aa94a4d9 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Took 0.06 seconds to destroy the instance on the hypervisor. [ 889.361213] env[61243]: DEBUG oslo.service.loopingcall [None req-e09ef140-7513-452f-829e-2105aa94a4d9 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.363954] env[61243]: DEBUG nova.compute.manager [-] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 889.364076] env[61243]: DEBUG nova.network.neutron [-] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 889.462336] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339025, 'name': CloneVM_Task, 'duration_secs': 1.576567} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.462639] env[61243]: INFO nova.virt.vmwareapi.vmops [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Created linked-clone VM from snapshot [ 889.463426] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d24c28f8-d46d-4051-8c48-afa3a240cda1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.482898] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Uploading image 2b9c3b0a-5f89-462e-99b2-0c88500ad15b {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 889.517121] env[61243]: DEBUG oslo_vmware.rw_handles [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 889.517121] env[61243]: value = "vm-285726" [ 889.517121] env[61243]: _type = "VirtualMachine" [ 889.517121] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 889.517463] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-45391408-99f6-41ce-afab-6ee62a303b18 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.525622] env[61243]: DEBUG oslo_vmware.rw_handles [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lease: (returnval){ [ 889.525622] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c6cb09-74e7-fc6c-923f-06089213cf5f" [ 889.525622] env[61243]: _type = "HttpNfcLease" [ 889.525622] env[61243]: } obtained for exporting VM: (result){ [ 889.525622] env[61243]: value = "vm-285726" [ 889.525622] env[61243]: _type = "VirtualMachine" [ 889.525622] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 889.526139] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the lease: (returnval){ [ 889.526139] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c6cb09-74e7-fc6c-923f-06089213cf5f" [ 889.526139] env[61243]: _type = "HttpNfcLease" [ 889.526139] env[61243]: } to be ready. {{(pid=61243) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 889.533643] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 889.533643] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c6cb09-74e7-fc6c-923f-06089213cf5f" [ 889.533643] env[61243]: _type = "HttpNfcLease" [ 889.533643] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 889.611050] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a5de984-facb-4170-b3d4-825e5d4e4df2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.619456] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dd4afe5-8cde-4ab3-849c-269ecf189fe0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.657157] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743e2169-b980-4279-b058-03e1e7b34929 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.668704] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527c7ae1-fa9f-7391-16a2-581878bb20b1, 'name': SearchDatastore_Task, 'duration_secs': 0.018868} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.671733] env[61243]: DEBUG oslo_concurrency.lockutils [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.672126] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 6036199d-cda9-42f8-8669-bc099d2b5ffe/6142e969-c114-4502-aa93-c018fb915a86-rescue.vmdk. {{(pid=61243) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 889.672706] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90e0cefa-328b-4534-b181-5497bfae513b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.676505] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8585c1f-96aa-40d9-b808-cf92b091a2df {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.694095] env[61243]: DEBUG nova.compute.provider_tree [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 889.697155] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 889.697155] env[61243]: value = "task-1339030" [ 889.697155] env[61243]: _type = "Task" [ 889.697155] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.707418] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339030, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.036174] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 890.036174] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c6cb09-74e7-fc6c-923f-06089213cf5f" [ 890.036174] env[61243]: _type = "HttpNfcLease" [ 890.036174] env[61243]: } is ready. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 890.036611] env[61243]: DEBUG oslo_vmware.rw_handles [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 890.036611] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c6cb09-74e7-fc6c-923f-06089213cf5f" [ 890.036611] env[61243]: _type = "HttpNfcLease" [ 890.036611] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 890.037614] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3334d78-6ca8-4c6c-b268-85ef437adcca {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.048356] env[61243]: DEBUG oslo_vmware.rw_handles [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52102955-2fe9-7e91-c906-0e215df20bae/disk-0.vmdk from lease info. {{(pid=61243) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 890.048558] env[61243]: DEBUG oslo_vmware.rw_handles [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52102955-2fe9-7e91-c906-0e215df20bae/disk-0.vmdk for reading. {{(pid=61243) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 890.050673] env[61243]: DEBUG nova.objects.instance [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lazy-loading 'flavor' on Instance uuid 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 890.162443] env[61243]: DEBUG nova.network.neutron [-] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.212372] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339030, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.218587] env[61243]: ERROR nova.scheduler.client.report [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [req-3a1ca63d-5ae5-4654-b243-36419a3e9dae] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3a1ca63d-5ae5-4654-b243-36419a3e9dae"}]} [ 890.238045] env[61243]: DEBUG nova.scheduler.client.report [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 890.253614] env[61243]: DEBUG nova.scheduler.client.report [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 890.253855] env[61243]: DEBUG nova.compute.provider_tree [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 890.268565] env[61243]: DEBUG nova.scheduler.client.report [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 890.293134] env[61243]: DEBUG nova.scheduler.client.report [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 890.333492] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-04c11dbd-c999-488d-8136-475bb4de39a6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.557264] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6744796f-0089-4daa-842d-d6e346cc91b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.883s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.579919] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a58e013-78b9-4025-8610-c3ca01187379 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.592889] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf5fbd3-4ab8-42ba-a7bd-ff217063e260 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.631106] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af8b30a-39a3-4230-b15e-b1e5e51f04f8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.642140] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657f779c-cd1f-4867-807a-a43fb6ed3c4f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.658828] env[61243]: DEBUG nova.compute.provider_tree [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.665124] env[61243]: INFO nova.compute.manager [-] [instance: ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5] Took 1.30 seconds to deallocate network for instance. [ 890.714157] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339030, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.678182} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.714472] env[61243]: INFO nova.virt.vmwareapi.ds_util [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 6036199d-cda9-42f8-8669-bc099d2b5ffe/6142e969-c114-4502-aa93-c018fb915a86-rescue.vmdk. [ 890.715663] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60b9a64d-9519-47b0-9da9-38abcaa9c211 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.742918] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 6036199d-cda9-42f8-8669-bc099d2b5ffe/6142e969-c114-4502-aa93-c018fb915a86-rescue.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.743566] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27673734-ee98-48a0-aa7a-e2843f539c5e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.764411] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 890.764411] env[61243]: value = "task-1339031" [ 890.764411] env[61243]: _type = "Task" [ 890.764411] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.774581] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339031, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.164535] env[61243]: DEBUG nova.scheduler.client.report [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.275816] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339031, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.673391] env[61243]: DEBUG oslo_concurrency.lockutils [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.886s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.675232] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.047s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.677504] env[61243]: INFO nova.compute.claims [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.695890] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e09ef140-7513-452f-829e-2105aa94a4d9 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "ae41ea9f-217e-4dbb-9278-bab6fa6c1ac5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.396s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.731837] env[61243]: INFO nova.scheduler.client.report [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleted allocations for instance ca230425-aaa7-4c32-81d3-379cbce0d41b [ 891.776551] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339031, 'name': ReconfigVM_Task, 'duration_secs': 0.556312} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.776851] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 6036199d-cda9-42f8-8669-bc099d2b5ffe/6142e969-c114-4502-aa93-c018fb915a86-rescue.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.777977] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5bc4acb-3c38-4a1d-a392-3ffe6dc10bc8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.812438] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87928618-2970-487d-9a81-f2a143a19aa0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.830454] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 891.830454] env[61243]: value = "task-1339032" [ 891.830454] env[61243]: _type = "Task" [ 891.830454] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.840978] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339032, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.242282] env[61243]: DEBUG oslo_concurrency.lockutils [None req-31ce4847-c5c4-43b0-a3b3-f2f7d684668c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "ca230425-aaa7-4c32-81d3-379cbce0d41b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.280s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.342537] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339032, 'name': ReconfigVM_Task, 'duration_secs': 0.325702} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.342887] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 892.343224] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-640cdbef-c5b6-44cf-9438-0d70fe634347 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.352101] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 892.352101] env[61243]: value = "task-1339033" [ 892.352101] env[61243]: _type = "Task" [ 892.352101] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.362506] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339033, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.524878] env[61243]: DEBUG nova.objects.instance [None req-b6a64894-abb8-4a53-b82d-94834e1f8d07 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Lazy-loading 'flavor' on Instance uuid 36db1ed5-846f-4ad6-8cee-38b73ff00321 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.667393] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "ebe8cc8e-cc35-4ec4-bed7-d55df72a2991" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.667647] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "ebe8cc8e-cc35-4ec4-bed7-d55df72a2991" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.837932] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "d46fb6dc-cdd9-4688-b5e7-9efde0f770c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.837932] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d46fb6dc-cdd9-4688-b5e7-9efde0f770c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.866715] env[61243]: DEBUG oslo_vmware.api [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339033, 'name': PowerOnVM_Task, 'duration_secs': 0.447156} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.867033] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 892.872531] env[61243]: DEBUG nova.compute.manager [None req-101a39b8-3671-40c5-b854-4cda47b181a6 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.873456] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecaeff45-f410-499f-857f-fd39ced5208e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.950977] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cab674c-e841-491e-b942-233296b5db9a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.959749] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d86f47-50cb-4f03-992f-674bb4b1f57a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.991510] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5339b92f-aa9a-4653-8408-3d1812043632 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.000601] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b7c1b1-9fb0-4e31-8cd4-40946904a235 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.015556] env[61243]: DEBUG nova.compute.provider_tree [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.030714] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b6a64894-abb8-4a53-b82d-94834e1f8d07 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Acquiring lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.030895] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b6a64894-abb8-4a53-b82d-94834e1f8d07 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Acquired lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.172635] env[61243]: DEBUG nova.compute.manager [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 893.344585] env[61243]: DEBUG nova.compute.manager [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 893.519157] env[61243]: DEBUG nova.scheduler.client.report [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.650616] env[61243]: DEBUG nova.network.neutron [None req-b6a64894-abb8-4a53-b82d-94834e1f8d07 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 893.681358] env[61243]: DEBUG nova.compute.manager [req-cf90a859-747a-4a7d-9415-82e5bfd45940 req-c3c4a76f-f513-46ed-84f7-4ce1c4af5405 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Received event network-changed-817c2755-c3a0-4de3-9427-322d92ba1d30 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 893.681358] env[61243]: DEBUG nova.compute.manager [req-cf90a859-747a-4a7d-9415-82e5bfd45940 req-c3c4a76f-f513-46ed-84f7-4ce1c4af5405 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Refreshing instance network info cache due to event network-changed-817c2755-c3a0-4de3-9427-322d92ba1d30. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 893.681358] env[61243]: DEBUG oslo_concurrency.lockutils [req-cf90a859-747a-4a7d-9415-82e5bfd45940 req-c3c4a76f-f513-46ed-84f7-4ce1c4af5405 service nova] Acquiring lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.695247] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.861062] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "5c063b4a-3fbb-4222-ab6d-7d27b6b99084" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.861062] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "5c063b4a-3fbb-4222-ab6d-7d27b6b99084" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.872277] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.023740] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.348s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.024374] env[61243]: DEBUG nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 894.027352] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.358s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.028952] env[61243]: INFO nova.compute.claims [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 894.364219] env[61243]: DEBUG nova.compute.manager [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 894.411595] env[61243]: DEBUG nova.network.neutron [None req-b6a64894-abb8-4a53-b82d-94834e1f8d07 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Updating instance_info_cache with network_info: [{"id": "817c2755-c3a0-4de3-9427-322d92ba1d30", "address": "fa:16:3e:5c:f9:55", "network": {"id": "f6dda51c-55ed-4b80-b54d-ade24626e5fd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1249632116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93625b58c8674dc0980e3d9fee62ab18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap817c2755-c3", "ovs_interfaceid": "817c2755-c3a0-4de3-9427-322d92ba1d30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.533414] env[61243]: DEBUG nova.compute.utils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 894.537549] env[61243]: DEBUG nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 894.537732] env[61243]: DEBUG nova.network.neutron [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 894.582115] env[61243]: DEBUG nova.policy [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18fd39290a104706a2e2633308d41213', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '39a655fc062e4322a5a129b66a45684f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 894.699090] env[61243]: INFO nova.compute.manager [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Unrescuing [ 894.699389] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "refresh_cache-6036199d-cda9-42f8-8669-bc099d2b5ffe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.699558] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquired lock "refresh_cache-6036199d-cda9-42f8-8669-bc099d2b5ffe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.699730] env[61243]: DEBUG nova.network.neutron [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 894.806993] env[61243]: DEBUG nova.objects.instance [None req-88227cf6-9573-47dc-a798-3485af6cf9f2 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Lazy-loading 'flavor' on Instance uuid 36db1ed5-846f-4ad6-8cee-38b73ff00321 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.866572] env[61243]: DEBUG nova.network.neutron [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Successfully created port: e579287f-20a4-48c1-a93d-3b88d6da8af0 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 894.893184] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.914993] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b6a64894-abb8-4a53-b82d-94834e1f8d07 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Releasing lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.914993] env[61243]: DEBUG nova.compute.manager [None req-b6a64894-abb8-4a53-b82d-94834e1f8d07 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Inject network info {{(pid=61243) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 894.915236] env[61243]: DEBUG nova.compute.manager [None req-b6a64894-abb8-4a53-b82d-94834e1f8d07 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] network_info to inject: |[{"id": "817c2755-c3a0-4de3-9427-322d92ba1d30", "address": "fa:16:3e:5c:f9:55", "network": {"id": "f6dda51c-55ed-4b80-b54d-ade24626e5fd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1249632116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93625b58c8674dc0980e3d9fee62ab18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap817c2755-c3", "ovs_interfaceid": "817c2755-c3a0-4de3-9427-322d92ba1d30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 894.920873] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b6a64894-abb8-4a53-b82d-94834e1f8d07 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Reconfiguring VM instance to set the machine id {{(pid=61243) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 894.920873] env[61243]: DEBUG oslo_concurrency.lockutils [req-cf90a859-747a-4a7d-9415-82e5bfd45940 req-c3c4a76f-f513-46ed-84f7-4ce1c4af5405 service nova] Acquired lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.920873] env[61243]: DEBUG nova.network.neutron [req-cf90a859-747a-4a7d-9415-82e5bfd45940 req-c3c4a76f-f513-46ed-84f7-4ce1c4af5405 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Refreshing network info cache for port 817c2755-c3a0-4de3-9427-322d92ba1d30 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 894.922950] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d31fa4ea-4b3f-4441-a404-922e94871058 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.943849] env[61243]: DEBUG oslo_vmware.api [None req-b6a64894-abb8-4a53-b82d-94834e1f8d07 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Waiting for the task: (returnval){ [ 894.943849] env[61243]: value = "task-1339034" [ 894.943849] env[61243]: _type = "Task" [ 894.943849] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.953899] env[61243]: DEBUG oslo_vmware.api [None req-b6a64894-abb8-4a53-b82d-94834e1f8d07 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1339034, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.038510] env[61243]: DEBUG nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 895.316748] env[61243]: DEBUG oslo_concurrency.lockutils [None req-88227cf6-9573-47dc-a798-3485af6cf9f2 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Acquiring lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.409160] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b124ed-d572-47c0-9c05-78614efe7c54 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.419294] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2cb55e-1785-4a0c-bffe-48b0b37b3d90 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.427948] env[61243]: DEBUG nova.network.neutron [req-cf90a859-747a-4a7d-9415-82e5bfd45940 req-c3c4a76f-f513-46ed-84f7-4ce1c4af5405 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Updated VIF entry in instance network info cache for port 817c2755-c3a0-4de3-9427-322d92ba1d30. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 895.427948] env[61243]: DEBUG nova.network.neutron [req-cf90a859-747a-4a7d-9415-82e5bfd45940 req-c3c4a76f-f513-46ed-84f7-4ce1c4af5405 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Updating instance_info_cache with network_info: [{"id": "817c2755-c3a0-4de3-9427-322d92ba1d30", "address": "fa:16:3e:5c:f9:55", "network": {"id": "f6dda51c-55ed-4b80-b54d-ade24626e5fd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1249632116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93625b58c8674dc0980e3d9fee62ab18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap817c2755-c3", "ovs_interfaceid": "817c2755-c3a0-4de3-9427-322d92ba1d30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.460617] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-251a2a85-5f7d-44d8-858f-2fba6d14e46f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.473948] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3568c4-1f67-40d8-a131-3c65c887ce5a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.480228] env[61243]: DEBUG oslo_vmware.api [None req-b6a64894-abb8-4a53-b82d-94834e1f8d07 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1339034, 'name': ReconfigVM_Task, 'duration_secs': 0.183627} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.480537] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b6a64894-abb8-4a53-b82d-94834e1f8d07 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Reconfigured VM instance to set the machine id {{(pid=61243) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 895.495292] env[61243]: DEBUG nova.compute.provider_tree [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.579025] env[61243]: DEBUG nova.network.neutron [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Updating instance_info_cache with network_info: [{"id": "15828714-3ad8-4f57-9307-10b1ac64495f", "address": "fa:16:3e:40:09:f1", "network": {"id": "8e2b6835-484b-4346-8bd1-034c53671110", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710127686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4fcef2317fdf47f8ac820aeef1765fef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15828714-3a", "ovs_interfaceid": "15828714-3ad8-4f57-9307-10b1ac64495f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.955485] env[61243]: DEBUG oslo_concurrency.lockutils [req-cf90a859-747a-4a7d-9415-82e5bfd45940 req-c3c4a76f-f513-46ed-84f7-4ce1c4af5405 service nova] Releasing lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.955928] env[61243]: DEBUG oslo_concurrency.lockutils [None req-88227cf6-9573-47dc-a798-3485af6cf9f2 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Acquired lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.000917] env[61243]: DEBUG nova.scheduler.client.report [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.050811] env[61243]: DEBUG nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 896.076868] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 896.077161] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 896.077331] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 896.077525] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 896.077679] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 896.077834] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 896.078056] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 896.078274] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 896.078466] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 896.078652] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 896.078845] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 896.079905] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4615b7-3902-4cbd-a4f9-81123af4600a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.082736] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Releasing lock "refresh_cache-6036199d-cda9-42f8-8669-bc099d2b5ffe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.083317] env[61243]: DEBUG nova.objects.instance [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lazy-loading 'flavor' on Instance uuid 6036199d-cda9-42f8-8669-bc099d2b5ffe {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.092200] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-974e8a5c-486d-4c5a-9928-db468ea0327d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.331882] env[61243]: DEBUG nova.network.neutron [None req-88227cf6-9573-47dc-a798-3485af6cf9f2 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 896.456427] env[61243]: DEBUG nova.compute.manager [req-409d40f9-d32e-451e-81c2-ce862f07e23d req-f9ae7e4e-1b35-4b7a-a31d-065381b13b63 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Received event network-changed-817c2755-c3a0-4de3-9427-322d92ba1d30 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.456627] env[61243]: DEBUG nova.compute.manager [req-409d40f9-d32e-451e-81c2-ce862f07e23d req-f9ae7e4e-1b35-4b7a-a31d-065381b13b63 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Refreshing instance network info cache due to event network-changed-817c2755-c3a0-4de3-9427-322d92ba1d30. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.456822] env[61243]: DEBUG oslo_concurrency.lockutils [req-409d40f9-d32e-451e-81c2-ce862f07e23d req-f9ae7e4e-1b35-4b7a-a31d-065381b13b63 service nova] Acquiring lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.506190] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.506728] env[61243]: DEBUG nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 896.509672] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.848s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.589915] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7017b3dd-4a1e-4301-adaa-9f6a333e4703 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.615969] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 896.616228] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-10dc5647-8af9-4e0b-b3fc-113a373a5bb8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.627022] env[61243]: DEBUG oslo_vmware.api [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 896.627022] env[61243]: value = "task-1339035" [ 896.627022] env[61243]: _type = "Task" [ 896.627022] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.635970] env[61243]: DEBUG oslo_vmware.api [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339035, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.757037] env[61243]: DEBUG nova.network.neutron [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Successfully updated port: e579287f-20a4-48c1-a93d-3b88d6da8af0 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 897.013300] env[61243]: DEBUG nova.compute.utils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 897.014847] env[61243]: DEBUG nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 897.015207] env[61243]: DEBUG nova.network.neutron [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 897.078305] env[61243]: DEBUG nova.compute.manager [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.079354] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e00f1c1-6ff7-4355-8f7b-28b0c49f7734 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.092871] env[61243]: DEBUG nova.policy [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18fd39290a104706a2e2633308d41213', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '39a655fc062e4322a5a129b66a45684f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 897.140376] env[61243]: DEBUG oslo_vmware.api [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339035, 'name': PowerOffVM_Task, 'duration_secs': 0.237926} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.140896] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 897.149066] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Reconfiguring VM instance instance-0000004d to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 897.149754] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-535f38a2-0d60-4d97-8aef-c80bb41e2866 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.170511] env[61243]: DEBUG nova.network.neutron [None req-88227cf6-9573-47dc-a798-3485af6cf9f2 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Updating instance_info_cache with network_info: [{"id": "817c2755-c3a0-4de3-9427-322d92ba1d30", "address": "fa:16:3e:5c:f9:55", "network": {"id": "f6dda51c-55ed-4b80-b54d-ade24626e5fd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1249632116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93625b58c8674dc0980e3d9fee62ab18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap817c2755-c3", "ovs_interfaceid": "817c2755-c3a0-4de3-9427-322d92ba1d30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.180139] env[61243]: DEBUG oslo_vmware.api [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 897.180139] env[61243]: value = "task-1339036" [ 897.180139] env[61243]: _type = "Task" [ 897.180139] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.194507] env[61243]: DEBUG oslo_vmware.api [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339036, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.260240] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "refresh_cache-f8816b86-10c8-4d64-9659-fbc301893d46" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.260469] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquired lock "refresh_cache-f8816b86-10c8-4d64-9659-fbc301893d46" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.260663] env[61243]: DEBUG nova.network.neutron [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 897.414785] env[61243]: DEBUG nova.network.neutron [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Successfully created port: f9010c84-231c-48c6-a2da-bcc9cd57ac3c {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 897.526168] env[61243]: DEBUG nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 897.557749] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 86179fb0-99df-4b10-a815-c19168e9521e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 897.557917] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 07ee984d-476e-484d-ba80-0ec2e411faa9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 897.558055] env[61243]: WARNING nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 897.558205] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 894eb798-6c9e-47cb-8eb5-4610fe184bc5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 897.558396] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 897.558602] env[61243]: WARNING nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 729d9bd9-1ab0-47bd-9b7a-11412d3dc608 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 897.558777] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 36db1ed5-846f-4ad6-8cee-38b73ff00321 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 897.558905] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 316572e0-c007-42cb-aaf0-3a8cfcaf24aa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 897.559699] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance c1a0d242-4eae-4c03-8341-840b41341f17 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 897.559936] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 897.560043] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 3f839aa5-2b9a-4807-b63b-931f74455532 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 897.560188] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 783d861c-8930-406e-8986-ef995de4c9dd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 897.560287] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 6036199d-cda9-42f8-8669-bc099d2b5ffe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 897.560405] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance f8816b86-10c8-4d64-9659-fbc301893d46 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 897.560519] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance d95a4d61-2bff-4d89-9ab8-f71e558f3409 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 897.591635] env[61243]: INFO nova.compute.manager [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] instance snapshotting [ 897.592684] env[61243]: DEBUG nova.objects.instance [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lazy-loading 'flavor' on Instance uuid 316572e0-c007-42cb-aaf0-3a8cfcaf24aa {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.674940] env[61243]: DEBUG oslo_concurrency.lockutils [None req-88227cf6-9573-47dc-a798-3485af6cf9f2 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Releasing lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.675219] env[61243]: DEBUG nova.compute.manager [None req-88227cf6-9573-47dc-a798-3485af6cf9f2 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Inject network info {{(pid=61243) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 897.675796] env[61243]: DEBUG nova.compute.manager [None req-88227cf6-9573-47dc-a798-3485af6cf9f2 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] network_info to inject: |[{"id": "817c2755-c3a0-4de3-9427-322d92ba1d30", "address": "fa:16:3e:5c:f9:55", "network": {"id": "f6dda51c-55ed-4b80-b54d-ade24626e5fd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1249632116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93625b58c8674dc0980e3d9fee62ab18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap817c2755-c3", "ovs_interfaceid": "817c2755-c3a0-4de3-9427-322d92ba1d30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 897.680231] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-88227cf6-9573-47dc-a798-3485af6cf9f2 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Reconfiguring VM instance to set the machine id {{(pid=61243) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 897.680575] env[61243]: DEBUG oslo_concurrency.lockutils [req-409d40f9-d32e-451e-81c2-ce862f07e23d req-f9ae7e4e-1b35-4b7a-a31d-065381b13b63 service nova] Acquired lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.680767] env[61243]: DEBUG nova.network.neutron [req-409d40f9-d32e-451e-81c2-ce862f07e23d req-f9ae7e4e-1b35-4b7a-a31d-065381b13b63 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Refreshing network info cache for port 817c2755-c3a0-4de3-9427-322d92ba1d30 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 897.682297] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43dafaaa-ecd4-4f07-83f1-a36be84ea5cb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.708025] env[61243]: DEBUG oslo_vmware.api [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339036, 'name': ReconfigVM_Task, 'duration_secs': 0.277087} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.709578] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Reconfigured VM instance instance-0000004d to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 897.709739] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 897.710067] env[61243]: DEBUG oslo_vmware.api [None req-88227cf6-9573-47dc-a798-3485af6cf9f2 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Waiting for the task: (returnval){ [ 897.710067] env[61243]: value = "task-1339037" [ 897.710067] env[61243]: _type = "Task" [ 897.710067] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.710310] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-20bfbf1d-e436-4dcc-9a76-3f70cc29ac91 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.723896] env[61243]: DEBUG oslo_vmware.api [None req-88227cf6-9573-47dc-a798-3485af6cf9f2 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1339037, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.725380] env[61243]: DEBUG oslo_vmware.api [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 897.725380] env[61243]: value = "task-1339038" [ 897.725380] env[61243]: _type = "Task" [ 897.725380] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.734918] env[61243]: DEBUG oslo_vmware.api [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339038, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.795840] env[61243]: DEBUG nova.network.neutron [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 897.966615] env[61243]: DEBUG nova.network.neutron [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Updating instance_info_cache with network_info: [{"id": "e579287f-20a4-48c1-a93d-3b88d6da8af0", "address": "fa:16:3e:0c:a4:bc", "network": {"id": "e10de432-a279-478d-a226-8cb1663f8f1e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-237966643-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39a655fc062e4322a5a129b66a45684f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape579287f-20", "ovs_interfaceid": "e579287f-20a4-48c1-a93d-3b88d6da8af0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.664022] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance e9a1c8af-7129-4366-bda1-1c3cf40eeab1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 898.677396] env[61243]: DEBUG nova.compute.manager [req-6dfc9922-6c9d-4815-b014-aa74f461d603 req-2196519a-cf94-4fdc-8dee-bb880f190562 service nova] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Received event network-vif-plugged-e579287f-20a4-48c1-a93d-3b88d6da8af0 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.677605] env[61243]: DEBUG oslo_concurrency.lockutils [req-6dfc9922-6c9d-4815-b014-aa74f461d603 req-2196519a-cf94-4fdc-8dee-bb880f190562 service nova] Acquiring lock "f8816b86-10c8-4d64-9659-fbc301893d46-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.677810] env[61243]: DEBUG oslo_concurrency.lockutils [req-6dfc9922-6c9d-4815-b014-aa74f461d603 req-2196519a-cf94-4fdc-8dee-bb880f190562 service nova] Lock "f8816b86-10c8-4d64-9659-fbc301893d46-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.677972] env[61243]: DEBUG oslo_concurrency.lockutils [req-6dfc9922-6c9d-4815-b014-aa74f461d603 req-2196519a-cf94-4fdc-8dee-bb880f190562 service nova] Lock "f8816b86-10c8-4d64-9659-fbc301893d46-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.678211] env[61243]: DEBUG nova.compute.manager [req-6dfc9922-6c9d-4815-b014-aa74f461d603 req-2196519a-cf94-4fdc-8dee-bb880f190562 service nova] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] No waiting events found dispatching network-vif-plugged-e579287f-20a4-48c1-a93d-3b88d6da8af0 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 898.678433] env[61243]: WARNING nova.compute.manager [req-6dfc9922-6c9d-4815-b014-aa74f461d603 req-2196519a-cf94-4fdc-8dee-bb880f190562 service nova] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Received unexpected event network-vif-plugged-e579287f-20a4-48c1-a93d-3b88d6da8af0 for instance with vm_state building and task_state spawning. [ 898.678605] env[61243]: DEBUG nova.compute.manager [req-6dfc9922-6c9d-4815-b014-aa74f461d603 req-2196519a-cf94-4fdc-8dee-bb880f190562 service nova] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Received event network-changed-e579287f-20a4-48c1-a93d-3b88d6da8af0 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.678760] env[61243]: DEBUG nova.compute.manager [req-6dfc9922-6c9d-4815-b014-aa74f461d603 req-2196519a-cf94-4fdc-8dee-bb880f190562 service nova] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Refreshing instance network info cache due to event network-changed-e579287f-20a4-48c1-a93d-3b88d6da8af0. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 898.678925] env[61243]: DEBUG oslo_concurrency.lockutils [req-6dfc9922-6c9d-4815-b014-aa74f461d603 req-2196519a-cf94-4fdc-8dee-bb880f190562 service nova] Acquiring lock "refresh_cache-f8816b86-10c8-4d64-9659-fbc301893d46" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.680967] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22f2ed10-e27b-45c0-915d-0f3a1d834bb6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.691754] env[61243]: DEBUG oslo_vmware.api [None req-88227cf6-9573-47dc-a798-3485af6cf9f2 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1339037, 'name': ReconfigVM_Task, 'duration_secs': 0.165076} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.710981] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-88227cf6-9573-47dc-a798-3485af6cf9f2 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Reconfigured VM instance to set the machine id {{(pid=61243) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 898.714329] env[61243]: DEBUG oslo_vmware.api [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339038, 'name': PowerOnVM_Task, 'duration_secs': 0.428227} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.716995] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b04cd95-6814-4963-aac8-27144f1e23ed {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.719564] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 898.719798] env[61243]: DEBUG nova.compute.manager [None req-5bfe25b6-70f2-429e-bccc-38aaec2d8b9f tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 898.722181] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e20838-b3d0-44d1-8ede-d4678ec56408 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.737700] env[61243]: DEBUG oslo_vmware.rw_handles [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52102955-2fe9-7e91-c906-0e215df20bae/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 898.739117] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4233a45f-fbdb-4179-a8ff-8a4adb00a176 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.746437] env[61243]: DEBUG oslo_vmware.rw_handles [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52102955-2fe9-7e91-c906-0e215df20bae/disk-0.vmdk is in state: ready. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 898.746437] env[61243]: ERROR oslo_vmware.rw_handles [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52102955-2fe9-7e91-c906-0e215df20bae/disk-0.vmdk due to incomplete transfer. [ 898.746889] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b4bdf612-62bd-459c-8440-da500f88120c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.755459] env[61243]: DEBUG oslo_vmware.rw_handles [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52102955-2fe9-7e91-c906-0e215df20bae/disk-0.vmdk. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 898.755809] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Uploaded image 2b9c3b0a-5f89-462e-99b2-0c88500ad15b to the Glance image server {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 898.758527] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Destroying the VM {{(pid=61243) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 898.758824] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-04ac8132-bd85-4483-a0fd-0535915a65f1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.766688] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 898.766688] env[61243]: value = "task-1339039" [ 898.766688] env[61243]: _type = "Task" [ 898.766688] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.776163] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339039, 'name': Destroy_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.069237] env[61243]: DEBUG nova.network.neutron [req-409d40f9-d32e-451e-81c2-ce862f07e23d req-f9ae7e4e-1b35-4b7a-a31d-065381b13b63 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Updated VIF entry in instance network info cache for port 817c2755-c3a0-4de3-9427-322d92ba1d30. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 899.069646] env[61243]: DEBUG nova.network.neutron [req-409d40f9-d32e-451e-81c2-ce862f07e23d req-f9ae7e4e-1b35-4b7a-a31d-065381b13b63 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Updating instance_info_cache with network_info: [{"id": "817c2755-c3a0-4de3-9427-322d92ba1d30", "address": "fa:16:3e:5c:f9:55", "network": {"id": "f6dda51c-55ed-4b80-b54d-ade24626e5fd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1249632116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93625b58c8674dc0980e3d9fee62ab18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap817c2755-c3", "ovs_interfaceid": "817c2755-c3a0-4de3-9427-322d92ba1d30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.130972] env[61243]: DEBUG nova.network.neutron [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Successfully updated port: f9010c84-231c-48c6-a2da-bcc9cd57ac3c {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 899.167841] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance ebe8cc8e-cc35-4ec4-bed7-d55df72a2991 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 899.171889] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Releasing lock "refresh_cache-f8816b86-10c8-4d64-9659-fbc301893d46" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.171889] env[61243]: DEBUG nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Instance network_info: |[{"id": "e579287f-20a4-48c1-a93d-3b88d6da8af0", "address": "fa:16:3e:0c:a4:bc", "network": {"id": "e10de432-a279-478d-a226-8cb1663f8f1e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-237966643-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39a655fc062e4322a5a129b66a45684f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape579287f-20", "ovs_interfaceid": "e579287f-20a4-48c1-a93d-3b88d6da8af0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 899.171889] env[61243]: DEBUG nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 899.172806] env[61243]: DEBUG oslo_concurrency.lockutils [req-6dfc9922-6c9d-4815-b014-aa74f461d603 req-2196519a-cf94-4fdc-8dee-bb880f190562 service nova] Acquired lock "refresh_cache-f8816b86-10c8-4d64-9659-fbc301893d46" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.172992] env[61243]: DEBUG nova.network.neutron [req-6dfc9922-6c9d-4815-b014-aa74f461d603 req-2196519a-cf94-4fdc-8dee-bb880f190562 service nova] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Refreshing network info cache for port e579287f-20a4-48c1-a93d-3b88d6da8af0 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 899.174036] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:a4:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73f6629b-7f80-4a5b-8f15-c7a1635b3c33', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e579287f-20a4-48c1-a93d-3b88d6da8af0', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 899.184087] env[61243]: DEBUG oslo.service.loopingcall [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.185241] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 899.185572] env[61243]: DEBUG oslo_concurrency.lockutils [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "c1a0d242-4eae-4c03-8341-840b41341f17" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.185837] env[61243]: DEBUG oslo_concurrency.lockutils [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "c1a0d242-4eae-4c03-8341-840b41341f17" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.185941] env[61243]: INFO nova.compute.manager [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Shelving [ 899.187215] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Acquiring lock "36db1ed5-846f-4ad6-8cee-38b73ff00321" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.188058] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Lock "36db1ed5-846f-4ad6-8cee-38b73ff00321" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.188058] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Acquiring lock "36db1ed5-846f-4ad6-8cee-38b73ff00321-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.188058] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Lock "36db1ed5-846f-4ad6-8cee-38b73ff00321-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.188058] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Lock "36db1ed5-846f-4ad6-8cee-38b73ff00321-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.189387] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9efc95fe-469c-4263-bd93-959385c11d60 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.207679] env[61243]: INFO nova.compute.manager [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Terminating instance [ 899.214518] env[61243]: DEBUG nova.compute.manager [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 899.214518] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 899.216621] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be017d64-8956-4c92-b394-fe9ddfa22f7e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.219940] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.222489] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-604c9ce7-d0a7-4e9c-b1ab-0769c3d6dbc2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.227359] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 899.227595] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 899.227755] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.227945] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 899.228116] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.228277] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 899.228519] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 899.228695] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 899.228865] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 899.229045] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 899.229224] env[61243]: DEBUG nova.virt.hardware [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.230233] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 899.230233] env[61243]: value = "task-1339040" [ 899.230233] env[61243]: _type = "Task" [ 899.230233] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.230986] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a793177-a895-49a1-8fdc-9b32a1749c69 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.236725] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.240157] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e655663-97d9-49aa-bd6d-e5f6098dc72f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.245617] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Creating Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 899.245964] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 899.245964] env[61243]: value = "task-1339041" [ 899.245964] env[61243]: _type = "Task" [ 899.245964] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.249648] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d0eeec99-2725-46a0-9e5c-194cf86ca47b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.257093] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46534921-2d9b-4cb6-91b4-5a36b952461d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.269030] env[61243]: DEBUG oslo_vmware.api [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Waiting for the task: (returnval){ [ 899.269030] env[61243]: value = "task-1339042" [ 899.269030] env[61243]: _type = "Task" [ 899.269030] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.269030] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339040, 'name': CreateVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.284989] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339041, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.285540] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 899.285540] env[61243]: value = "task-1339043" [ 899.285540] env[61243]: _type = "Task" [ 899.285540] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.304075] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339039, 'name': Destroy_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.304587] env[61243]: DEBUG oslo_vmware.api [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1339042, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.309931] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339043, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.572616] env[61243]: DEBUG oslo_concurrency.lockutils [req-409d40f9-d32e-451e-81c2-ce862f07e23d req-f9ae7e4e-1b35-4b7a-a31d-065381b13b63 service nova] Releasing lock "refresh_cache-36db1ed5-846f-4ad6-8cee-38b73ff00321" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.634891] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "refresh_cache-d95a4d61-2bff-4d89-9ab8-f71e558f3409" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.635140] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquired lock "refresh_cache-d95a4d61-2bff-4d89-9ab8-f71e558f3409" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.635370] env[61243]: DEBUG nova.network.neutron [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 899.674397] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance d46fb6dc-cdd9-4688-b5e7-9efde0f770c9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 899.750101] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339040, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.768824] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339041, 'name': PowerOffVM_Task, 'duration_secs': 0.282072} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.769379] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 899.774341] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9ebe72-fffd-427d-8b09-002219de146f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.810815] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339039, 'name': Destroy_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.816240] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-695f2a3b-c08e-4a84-bbf9-b9699820e953 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.822459] env[61243]: DEBUG oslo_vmware.api [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1339042, 'name': PowerOffVM_Task, 'duration_secs': 0.270098} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.823488] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 899.823647] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 899.823923] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b12a9ed-515b-4075-83a2-f24cc8dc1b21 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.832542] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "6036199d-cda9-42f8-8669-bc099d2b5ffe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.832770] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "6036199d-cda9-42f8-8669-bc099d2b5ffe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.832972] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "6036199d-cda9-42f8-8669-bc099d2b5ffe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.833170] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "6036199d-cda9-42f8-8669-bc099d2b5ffe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.833343] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "6036199d-cda9-42f8-8669-bc099d2b5ffe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.834853] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339043, 'name': CreateSnapshot_Task, 'duration_secs': 0.497746} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.837598] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Created Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 899.838089] env[61243]: INFO nova.compute.manager [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Terminating instance [ 899.840151] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4381a3ca-1c9f-48dd-b80e-5ab4eecb724b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.843289] env[61243]: DEBUG nova.compute.manager [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 899.843289] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 899.843867] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a679ee2-2762-46f2-8c06-9a9d3084e9e5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.857779] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.858333] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e692ee0b-c072-48d4-b7a3-5cef24015c38 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.866104] env[61243]: DEBUG oslo_vmware.api [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 899.866104] env[61243]: value = "task-1339045" [ 899.866104] env[61243]: _type = "Task" [ 899.866104] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.878936] env[61243]: DEBUG oslo_vmware.api [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339045, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.920344] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 899.920344] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 899.920964] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Deleting the datastore file [datastore1] 36db1ed5-846f-4ad6-8cee-38b73ff00321 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 899.920964] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-77d64538-12f7-482d-b748-69d3534e93d8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.929086] env[61243]: DEBUG oslo_vmware.api [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Waiting for the task: (returnval){ [ 899.929086] env[61243]: value = "task-1339046" [ 899.929086] env[61243]: _type = "Task" [ 899.929086] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.937525] env[61243]: DEBUG oslo_concurrency.lockutils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "09587d68-f7fd-4648-bd5c-e53b1b6464e8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.937783] env[61243]: DEBUG oslo_concurrency.lockutils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "09587d68-f7fd-4648-bd5c-e53b1b6464e8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.942756] env[61243]: DEBUG oslo_vmware.api [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1339046, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.953445] env[61243]: DEBUG nova.network.neutron [req-6dfc9922-6c9d-4815-b014-aa74f461d603 req-2196519a-cf94-4fdc-8dee-bb880f190562 service nova] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Updated VIF entry in instance network info cache for port e579287f-20a4-48c1-a93d-3b88d6da8af0. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 899.953974] env[61243]: DEBUG nova.network.neutron [req-6dfc9922-6c9d-4815-b014-aa74f461d603 req-2196519a-cf94-4fdc-8dee-bb880f190562 service nova] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Updating instance_info_cache with network_info: [{"id": "e579287f-20a4-48c1-a93d-3b88d6da8af0", "address": "fa:16:3e:0c:a4:bc", "network": {"id": "e10de432-a279-478d-a226-8cb1663f8f1e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-237966643-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39a655fc062e4322a5a129b66a45684f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape579287f-20", "ovs_interfaceid": "e579287f-20a4-48c1-a93d-3b88d6da8af0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.169087] env[61243]: DEBUG nova.network.neutron [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 900.177052] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 5c063b4a-3fbb-4222-ab6d-7d27b6b99084 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 900.177408] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 900.177577] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3008MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 900.249146] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339040, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.281501] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339039, 'name': Destroy_Task, 'duration_secs': 1.358434} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.284286] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Destroyed the VM [ 900.284538] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Deleting Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 900.286922] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-852d2c0a-37f9-40ed-b5e7-81d8171e54eb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.297076] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 900.297076] env[61243]: value = "task-1339047" [ 900.297076] env[61243]: _type = "Task" [ 900.297076] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.311189] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339047, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.330150] env[61243]: DEBUG nova.network.neutron [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Updating instance_info_cache with network_info: [{"id": "f9010c84-231c-48c6-a2da-bcc9cd57ac3c", "address": "fa:16:3e:6e:50:d0", "network": {"id": "e10de432-a279-478d-a226-8cb1663f8f1e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-237966643-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39a655fc062e4322a5a129b66a45684f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9010c84-23", "ovs_interfaceid": "f9010c84-231c-48c6-a2da-bcc9cd57ac3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.338703] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Creating Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 900.339161] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0e026716-fed9-41fa-86c3-1ef0a19ca70e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.347929] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 900.347929] env[61243]: value = "task-1339048" [ 900.347929] env[61243]: _type = "Task" [ 900.347929] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.358083] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339048, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.368095] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Creating linked-clone VM from snapshot {{(pid=61243) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 900.368988] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-28cbb773-1a54-416a-ab73-cd5de7842150 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.380448] env[61243]: DEBUG oslo_vmware.api [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339045, 'name': PowerOffVM_Task, 'duration_secs': 0.205024} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.384555] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 900.384555] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 900.384843] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 900.384843] env[61243]: value = "task-1339049" [ 900.384843] env[61243]: _type = "Task" [ 900.384843] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.384966] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7882cbf5-468e-4ec8-b596-4643e001aad8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.396044] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339049, 'name': CloneVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.441852] env[61243]: DEBUG nova.compute.manager [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 900.445539] env[61243]: DEBUG oslo_vmware.api [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Task: {'id': task-1339046, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167391} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.446037] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 900.446235] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 900.446443] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 900.447016] env[61243]: INFO nova.compute.manager [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Took 1.23 seconds to destroy the instance on the hypervisor. [ 900.447016] env[61243]: DEBUG oslo.service.loopingcall [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.447175] env[61243]: DEBUG nova.compute.manager [-] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 900.447175] env[61243]: DEBUG nova.network.neutron [-] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 900.457059] env[61243]: DEBUG oslo_concurrency.lockutils [req-6dfc9922-6c9d-4815-b014-aa74f461d603 req-2196519a-cf94-4fdc-8dee-bb880f190562 service nova] Releasing lock "refresh_cache-f8816b86-10c8-4d64-9659-fbc301893d46" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.461419] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 900.461693] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 900.461941] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Deleting the datastore file [datastore2] 6036199d-cda9-42f8-8669-bc099d2b5ffe {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.462234] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-553c22e7-ef2a-496c-9cb6-dc4208d85c98 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.472120] env[61243]: DEBUG oslo_vmware.api [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 900.472120] env[61243]: value = "task-1339051" [ 900.472120] env[61243]: _type = "Task" [ 900.472120] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.486657] env[61243]: DEBUG oslo_vmware.api [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339051, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.500022] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf84363b-6c0d-4e2b-862d-1c3c0e610ef5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.509112] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4730d781-02cc-4750-8e51-0bf300457813 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.546088] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd50d5d-d817-47ce-9aa3-84defdc76da8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.556358] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b64d60-abd1-4327-b8b3-bf1cfb9ad09c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.574276] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 900.733308] env[61243]: DEBUG nova.compute.manager [req-4d86b9fb-d34c-4a86-979c-101bcc8f6a17 req-bfd90733-03c1-4f16-bb58-35d4ce208723 service nova] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Received event network-vif-plugged-f9010c84-231c-48c6-a2da-bcc9cd57ac3c {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.733577] env[61243]: DEBUG oslo_concurrency.lockutils [req-4d86b9fb-d34c-4a86-979c-101bcc8f6a17 req-bfd90733-03c1-4f16-bb58-35d4ce208723 service nova] Acquiring lock "d95a4d61-2bff-4d89-9ab8-f71e558f3409-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.733778] env[61243]: DEBUG oslo_concurrency.lockutils [req-4d86b9fb-d34c-4a86-979c-101bcc8f6a17 req-bfd90733-03c1-4f16-bb58-35d4ce208723 service nova] Lock "d95a4d61-2bff-4d89-9ab8-f71e558f3409-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.734015] env[61243]: DEBUG oslo_concurrency.lockutils [req-4d86b9fb-d34c-4a86-979c-101bcc8f6a17 req-bfd90733-03c1-4f16-bb58-35d4ce208723 service nova] Lock "d95a4d61-2bff-4d89-9ab8-f71e558f3409-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.734323] env[61243]: DEBUG nova.compute.manager [req-4d86b9fb-d34c-4a86-979c-101bcc8f6a17 req-bfd90733-03c1-4f16-bb58-35d4ce208723 service nova] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] No waiting events found dispatching network-vif-plugged-f9010c84-231c-48c6-a2da-bcc9cd57ac3c {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 900.734524] env[61243]: WARNING nova.compute.manager [req-4d86b9fb-d34c-4a86-979c-101bcc8f6a17 req-bfd90733-03c1-4f16-bb58-35d4ce208723 service nova] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Received unexpected event network-vif-plugged-f9010c84-231c-48c6-a2da-bcc9cd57ac3c for instance with vm_state building and task_state spawning. [ 900.734711] env[61243]: DEBUG nova.compute.manager [req-4d86b9fb-d34c-4a86-979c-101bcc8f6a17 req-bfd90733-03c1-4f16-bb58-35d4ce208723 service nova] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Received event network-changed-f9010c84-231c-48c6-a2da-bcc9cd57ac3c {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.734900] env[61243]: DEBUG nova.compute.manager [req-4d86b9fb-d34c-4a86-979c-101bcc8f6a17 req-bfd90733-03c1-4f16-bb58-35d4ce208723 service nova] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Refreshing instance network info cache due to event network-changed-f9010c84-231c-48c6-a2da-bcc9cd57ac3c. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 900.735106] env[61243]: DEBUG oslo_concurrency.lockutils [req-4d86b9fb-d34c-4a86-979c-101bcc8f6a17 req-bfd90733-03c1-4f16-bb58-35d4ce208723 service nova] Acquiring lock "refresh_cache-d95a4d61-2bff-4d89-9ab8-f71e558f3409" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.752022] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339040, 'name': CreateVM_Task, 'duration_secs': 1.16199} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.752022] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 900.752022] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.752022] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.752022] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 900.752022] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a37d895-e4be-42f2-adab-7b82bacdf393 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.759543] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 900.759543] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5230ac8c-54c1-8971-96ce-41933045d375" [ 900.759543] env[61243]: _type = "Task" [ 900.759543] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.773965] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5230ac8c-54c1-8971-96ce-41933045d375, 'name': SearchDatastore_Task, 'duration_secs': 0.010878} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.773965] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.773965] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 900.775639] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.775639] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.775639] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 900.775639] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97cd8a4c-3bfd-4da7-ae51-f722ed9e5e9b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.787331] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 900.787566] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 900.788502] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6498e7cf-6a1e-4b02-ac33-b99790097e56 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.799502] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 900.799502] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523921e9-7ca8-0e69-308e-ef6b0974b902" [ 900.799502] env[61243]: _type = "Task" [ 900.799502] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.813232] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523921e9-7ca8-0e69-308e-ef6b0974b902, 'name': SearchDatastore_Task, 'duration_secs': 0.010976} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.817160] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339047, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.817424] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1959fb47-1dfc-413a-b738-83481e06d058 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.823946] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 900.823946] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528642bd-25e4-e222-42ee-1f5d551610c7" [ 900.823946] env[61243]: _type = "Task" [ 900.823946] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.833982] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Releasing lock "refresh_cache-d95a4d61-2bff-4d89-9ab8-f71e558f3409" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.834365] env[61243]: DEBUG nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Instance network_info: |[{"id": "f9010c84-231c-48c6-a2da-bcc9cd57ac3c", "address": "fa:16:3e:6e:50:d0", "network": {"id": "e10de432-a279-478d-a226-8cb1663f8f1e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-237966643-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39a655fc062e4322a5a129b66a45684f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9010c84-23", "ovs_interfaceid": "f9010c84-231c-48c6-a2da-bcc9cd57ac3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 900.834723] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528642bd-25e4-e222-42ee-1f5d551610c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.835033] env[61243]: DEBUG oslo_concurrency.lockutils [req-4d86b9fb-d34c-4a86-979c-101bcc8f6a17 req-bfd90733-03c1-4f16-bb58-35d4ce208723 service nova] Acquired lock "refresh_cache-d95a4d61-2bff-4d89-9ab8-f71e558f3409" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.835254] env[61243]: DEBUG nova.network.neutron [req-4d86b9fb-d34c-4a86-979c-101bcc8f6a17 req-bfd90733-03c1-4f16-bb58-35d4ce208723 service nova] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Refreshing network info cache for port f9010c84-231c-48c6-a2da-bcc9cd57ac3c {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 900.836638] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:50:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73f6629b-7f80-4a5b-8f15-c7a1635b3c33', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f9010c84-231c-48c6-a2da-bcc9cd57ac3c', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 900.845337] env[61243]: DEBUG oslo.service.loopingcall [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.846471] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 900.846766] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-69c1973b-5100-47a0-a0b3-08c478b8c376 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.874042] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339048, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.875597] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 900.875597] env[61243]: value = "task-1339052" [ 900.875597] env[61243]: _type = "Task" [ 900.875597] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.884852] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339052, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.899041] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339049, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.965978] env[61243]: DEBUG oslo_concurrency.lockutils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.981899] env[61243]: DEBUG oslo_vmware.api [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339051, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.23994} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.981899] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 900.982175] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 900.982515] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 900.982617] env[61243]: INFO nova.compute.manager [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Took 1.14 seconds to destroy the instance on the hypervisor. [ 900.982846] env[61243]: DEBUG oslo.service.loopingcall [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.986020] env[61243]: DEBUG nova.compute.manager [-] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 900.986020] env[61243]: DEBUG nova.network.neutron [-] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 901.100424] env[61243]: ERROR nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [req-d7d322f6-2c02-4f9c-a2c6-8a808d631c22] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d7d322f6-2c02-4f9c-a2c6-8a808d631c22"}]} [ 901.117883] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 901.133034] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 901.133475] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 901.154485] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 901.178864] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 901.327454] env[61243]: DEBUG oslo_vmware.api [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339047, 'name': RemoveSnapshot_Task, 'duration_secs': 0.68743} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.327454] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Deleted Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 901.327454] env[61243]: INFO nova.compute.manager [None req-338c4645-cc4e-40f5-8319-a019c27b164c tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Took 16.49 seconds to snapshot the instance on the hypervisor. [ 901.343732] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528642bd-25e4-e222-42ee-1f5d551610c7, 'name': SearchDatastore_Task, 'duration_secs': 0.011287} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.346589] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.346943] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] f8816b86-10c8-4d64-9659-fbc301893d46/f8816b86-10c8-4d64-9659-fbc301893d46.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 901.349406] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a557be1d-70b7-4651-9cae-f6beb0c9d919 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.359154] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 901.359154] env[61243]: value = "task-1339053" [ 901.359154] env[61243]: _type = "Task" [ 901.359154] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.371513] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339053, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.377369] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339048, 'name': CreateSnapshot_Task, 'duration_secs': 0.606453} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.381547] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Created Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 901.382750] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3fda03d-59e1-4017-897e-589dbf553e22 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.401178] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339052, 'name': CreateVM_Task, 'duration_secs': 0.403621} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.404041] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.405034] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.405731] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.406134] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.410251] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0371f14-cd15-4cd4-b417-ff2bc3511068 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.412857] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339049, 'name': CloneVM_Task} progress is 95%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.420631] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 901.420631] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527d75df-43e4-a704-14a6-d1a7f1d5bed3" [ 901.420631] env[61243]: _type = "Task" [ 901.420631] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.430895] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527d75df-43e4-a704-14a6-d1a7f1d5bed3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.544446] env[61243]: DEBUG nova.network.neutron [-] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.567303] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10652f9c-d353-4425-8d25-55164910a593 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.580267] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0089de-ff7f-4a09-9a6d-1c9f7bc03ba9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.614654] env[61243]: DEBUG nova.network.neutron [req-4d86b9fb-d34c-4a86-979c-101bcc8f6a17 req-bfd90733-03c1-4f16-bb58-35d4ce208723 service nova] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Updated VIF entry in instance network info cache for port f9010c84-231c-48c6-a2da-bcc9cd57ac3c. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 901.615055] env[61243]: DEBUG nova.network.neutron [req-4d86b9fb-d34c-4a86-979c-101bcc8f6a17 req-bfd90733-03c1-4f16-bb58-35d4ce208723 service nova] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Updating instance_info_cache with network_info: [{"id": "f9010c84-231c-48c6-a2da-bcc9cd57ac3c", "address": "fa:16:3e:6e:50:d0", "network": {"id": "e10de432-a279-478d-a226-8cb1663f8f1e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-237966643-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39a655fc062e4322a5a129b66a45684f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9010c84-23", "ovs_interfaceid": "f9010c84-231c-48c6-a2da-bcc9cd57ac3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.617091] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4b6deb5-7f0e-4efc-a01c-23693506884f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.625726] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa9e380-3854-45cc-9974-2ad412147179 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.640741] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 901.814875] env[61243]: DEBUG nova.network.neutron [-] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.874014] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339053, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.898042] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339049, 'name': CloneVM_Task, 'duration_secs': 1.321079} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.898331] env[61243]: INFO nova.virt.vmwareapi.vmops [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Created linked-clone VM from snapshot [ 901.901046] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43051fb2-cf74-4abb-9574-3c9cd328bac1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.907581] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Uploading image 6fe66989-aad8-47fb-9027-b19aee66207e {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 901.921310] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Creating linked-clone VM from snapshot {{(pid=61243) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 901.921664] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-10693275-abfa-40da-8558-196fe0cfab26 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.936301] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527d75df-43e4-a704-14a6-d1a7f1d5bed3, 'name': SearchDatastore_Task, 'duration_secs': 0.013076} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.938166] env[61243]: DEBUG oslo_vmware.rw_handles [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 901.938166] env[61243]: value = "vm-285729" [ 901.938166] env[61243]: _type = "VirtualMachine" [ 901.938166] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 901.939748] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.939998] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 901.940260] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.940418] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.940601] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 901.941353] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-159b0b9c-688d-4796-863d-16b21fa56f8a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.942420] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 901.942420] env[61243]: value = "task-1339054" [ 901.942420] env[61243]: _type = "Task" [ 901.942420] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.942720] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9d66a14e-2df8-4909-9ef4-daea952e09e1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.959668] env[61243]: DEBUG oslo_vmware.rw_handles [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lease: (returnval){ [ 901.959668] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52898607-f3e8-9779-63c7-c02212de0865" [ 901.959668] env[61243]: _type = "HttpNfcLease" [ 901.959668] env[61243]: } obtained for exporting VM: (result){ [ 901.959668] env[61243]: value = "vm-285729" [ 901.959668] env[61243]: _type = "VirtualMachine" [ 901.959668] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 901.961099] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the lease: (returnval){ [ 901.961099] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52898607-f3e8-9779-63c7-c02212de0865" [ 901.961099] env[61243]: _type = "HttpNfcLease" [ 901.961099] env[61243]: } to be ready. {{(pid=61243) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 901.969101] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339054, 'name': CloneVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.976016] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 901.976016] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52898607-f3e8-9779-63c7-c02212de0865" [ 901.976016] env[61243]: _type = "HttpNfcLease" [ 901.976016] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 901.977036] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.977036] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 901.977910] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-869f6a5e-8c4b-446a-97e8-5fd03392ec77 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.986548] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 901.986548] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521d32d5-9d86-2127-a762-2d2cd06189ea" [ 901.986548] env[61243]: _type = "Task" [ 901.986548] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.993392] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521d32d5-9d86-2127-a762-2d2cd06189ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.050964] env[61243]: INFO nova.compute.manager [-] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Took 1.60 seconds to deallocate network for instance. [ 902.121393] env[61243]: DEBUG oslo_concurrency.lockutils [req-4d86b9fb-d34c-4a86-979c-101bcc8f6a17 req-bfd90733-03c1-4f16-bb58-35d4ce208723 service nova] Releasing lock "refresh_cache-d95a4d61-2bff-4d89-9ab8-f71e558f3409" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.177614] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 104 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 902.177913] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 104 to 105 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 902.178088] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 902.317921] env[61243]: INFO nova.compute.manager [-] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Took 1.33 seconds to deallocate network for instance. [ 902.371516] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339053, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53905} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.371842] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] f8816b86-10c8-4d64-9659-fbc301893d46/f8816b86-10c8-4d64-9659-fbc301893d46.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 902.372106] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 902.372388] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3690386a-4e28-4e16-a76c-0d7594634b30 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.379240] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 902.379240] env[61243]: value = "task-1339056" [ 902.379240] env[61243]: _type = "Task" [ 902.379240] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.386982] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339056, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.454801] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339054, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.467855] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 902.467855] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52898607-f3e8-9779-63c7-c02212de0865" [ 902.467855] env[61243]: _type = "HttpNfcLease" [ 902.467855] env[61243]: } is ready. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 902.468193] env[61243]: DEBUG oslo_vmware.rw_handles [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 902.468193] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52898607-f3e8-9779-63c7-c02212de0865" [ 902.468193] env[61243]: _type = "HttpNfcLease" [ 902.468193] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 902.468905] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba788d9-c89a-4ff1-84f6-9b2598d24110 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.476672] env[61243]: DEBUG oslo_vmware.rw_handles [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524777f6-6cea-7f74-59c6-034ba8434556/disk-0.vmdk from lease info. {{(pid=61243) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 902.476854] env[61243]: DEBUG oslo_vmware.rw_handles [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524777f6-6cea-7f74-59c6-034ba8434556/disk-0.vmdk for reading. {{(pid=61243) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 902.546484] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521d32d5-9d86-2127-a762-2d2cd06189ea, 'name': SearchDatastore_Task, 'duration_secs': 0.012777} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.547304] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94c92565-3a80-4cff-98ef-5a9023c3f9d6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.552997] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 902.552997] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a622d1-2130-4a0e-5b9a-3c0fcc8e07a9" [ 902.552997] env[61243]: _type = "Task" [ 902.552997] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.558270] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.561710] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a622d1-2130-4a0e-5b9a-3c0fcc8e07a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.576771] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e8d420b2-5092-4310-b05f-75b48c4803f7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.683594] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61243) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 902.683968] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.174s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.684278] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.300s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.684555] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.686643] env[61243]: DEBUG oslo_concurrency.lockutils [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.728s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.686856] env[61243]: DEBUG oslo_concurrency.lockutils [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.689159] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.364s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.690676] env[61243]: INFO nova.compute.claims [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 902.720156] env[61243]: INFO nova.scheduler.client.report [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Deleted allocations for instance 729d9bd9-1ab0-47bd-9b7a-11412d3dc608 [ 902.722282] env[61243]: INFO nova.scheduler.client.report [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Deleted allocations for instance 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00 [ 902.770324] env[61243]: DEBUG nova.compute.manager [req-30c54e94-8e2d-4cdb-8f0e-cbfe6ee60156 req-777f58c0-5835-4d74-8739-9aec59636091 service nova] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Received event network-vif-deleted-817c2755-c3a0-4de3-9427-322d92ba1d30 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.770324] env[61243]: DEBUG nova.compute.manager [req-30c54e94-8e2d-4cdb-8f0e-cbfe6ee60156 req-777f58c0-5835-4d74-8739-9aec59636091 service nova] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Received event network-vif-deleted-15828714-3ad8-4f57-9307-10b1ac64495f {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.825443] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.891478] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339056, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072525} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.891923] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 902.892777] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0439386-f573-4838-beb1-8b0bd439f03c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.917254] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] f8816b86-10c8-4d64-9659-fbc301893d46/f8816b86-10c8-4d64-9659-fbc301893d46.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.917541] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-510e31b1-5eac-4096-a7d0-9f6418f2cd6f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.939964] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 902.939964] env[61243]: value = "task-1339057" [ 902.939964] env[61243]: _type = "Task" [ 902.939964] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.952279] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339057, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.958949] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339054, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.066210] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a622d1-2130-4a0e-5b9a-3c0fcc8e07a9, 'name': SearchDatastore_Task, 'duration_secs': 0.010254} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.066941] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.067130] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] d95a4d61-2bff-4d89-9ab8-f71e558f3409/d95a4d61-2bff-4d89-9ab8-f71e558f3409.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.067508] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5eda1406-5ef1-4b36-baee-4dc2ef22ad64 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.076572] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 903.076572] env[61243]: value = "task-1339058" [ 903.076572] env[61243]: _type = "Task" [ 903.076572] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.086745] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339058, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.237123] env[61243]: DEBUG oslo_concurrency.lockutils [None req-55f9e60b-e613-4bf8-99e0-c42a7855728e tempest-ServerShowV247Test-969883687 tempest-ServerShowV247Test-969883687-project-member] Lock "729d9bd9-1ab0-47bd-9b7a-11412d3dc608" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.032s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.239043] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6c27569-ccd7-4b51-84de-22f58eb3d6a1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.090s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.454722] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339057, 'name': ReconfigVM_Task, 'duration_secs': 0.326881} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.459720] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Reconfigured VM instance instance-0000004e to attach disk [datastore2] f8816b86-10c8-4d64-9659-fbc301893d46/f8816b86-10c8-4d64-9659-fbc301893d46.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 903.460885] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-85b3cb98-b782-4d46-8777-26cca7f2043b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.467643] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339054, 'name': CloneVM_Task} progress is 95%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.478069] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 903.478069] env[61243]: value = "task-1339059" [ 903.478069] env[61243]: _type = "Task" [ 903.478069] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.491899] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339059, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.588673] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339058, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.652576] env[61243]: DEBUG oslo_concurrency.lockutils [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "86179fb0-99df-4b10-a815-c19168e9521e" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.652928] env[61243]: DEBUG oslo_concurrency.lockutils [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "86179fb0-99df-4b10-a815-c19168e9521e" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.653172] env[61243]: INFO nova.compute.manager [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Shelving [ 903.962513] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339054, 'name': CloneVM_Task, 'duration_secs': 1.651005} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.962899] env[61243]: INFO nova.virt.vmwareapi.vmops [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Created linked-clone VM from snapshot [ 903.963915] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71cd1afa-dfff-4234-a7a3-8a9eaddd412f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.968410] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a37731-cc72-4e88-baf3-028dfdf6cd29 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.978164] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Uploading image 9a127f45-caa9-4072-8a3a-d6944f9047b8 {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 903.987741] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f482f6-cb69-43bf-af64-40c8e6aab9ab {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.997125] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339059, 'name': Rename_Task, 'duration_secs': 0.244677} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.028423] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 904.031355] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-343af1cd-81ec-447d-9fc0-5802a50c1a34 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.034252] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295b2cd5-9802-45f7-8a4b-257c4b1031a1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.043950] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72140f4f-55f6-4d98-b549-6882cef4934d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.050474] env[61243]: DEBUG oslo_vmware.rw_handles [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 904.050474] env[61243]: value = "vm-285732" [ 904.050474] env[61243]: _type = "VirtualMachine" [ 904.050474] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 904.051494] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f3a07a10-a426-4aee-a6f5-1ca7f8e5102e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.065275] env[61243]: DEBUG nova.compute.provider_tree [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 904.075658] env[61243]: DEBUG oslo_vmware.rw_handles [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lease: (returnval){ [ 904.075658] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e12582-6774-3df5-20d6-cd6e7fc2e108" [ 904.075658] env[61243]: _type = "HttpNfcLease" [ 904.075658] env[61243]: } obtained for exporting VM: (result){ [ 904.075658] env[61243]: value = "vm-285732" [ 904.075658] env[61243]: _type = "VirtualMachine" [ 904.075658] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 904.076122] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the lease: (returnval){ [ 904.076122] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e12582-6774-3df5-20d6-cd6e7fc2e108" [ 904.076122] env[61243]: _type = "HttpNfcLease" [ 904.076122] env[61243]: } to be ready. {{(pid=61243) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 904.076429] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 904.076429] env[61243]: value = "task-1339060" [ 904.076429] env[61243]: _type = "Task" [ 904.076429] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.096634] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339058, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.543536} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.097015] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 904.097015] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e12582-6774-3df5-20d6-cd6e7fc2e108" [ 904.097015] env[61243]: _type = "HttpNfcLease" [ 904.097015] env[61243]: } is ready. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 904.100492] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] d95a4d61-2bff-4d89-9ab8-f71e558f3409/d95a4d61-2bff-4d89-9ab8-f71e558f3409.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 904.100908] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 904.101339] env[61243]: DEBUG oslo_vmware.rw_handles [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 904.101339] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e12582-6774-3df5-20d6-cd6e7fc2e108" [ 904.101339] env[61243]: _type = "HttpNfcLease" [ 904.101339] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 904.101771] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339060, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.102017] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-adb50407-b1ec-46b8-af6c-a2812598345a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.104837] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f9acb6-ee70-474e-92fa-75554b2a139d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.116370] env[61243]: DEBUG oslo_vmware.rw_handles [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52fd5ec2-92bb-9b45-a78f-c49232bad080/disk-0.vmdk from lease info. {{(pid=61243) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 904.116716] env[61243]: DEBUG oslo_vmware.rw_handles [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52fd5ec2-92bb-9b45-a78f-c49232bad080/disk-0.vmdk for reading. {{(pid=61243) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 904.119571] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 904.119571] env[61243]: value = "task-1339062" [ 904.119571] env[61243]: _type = "Task" [ 904.119571] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.193580] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339062, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.194243] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 904.194957] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-522cd836-7dad-4de5-82bb-f136bafd843a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.203884] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 904.203884] env[61243]: value = "task-1339063" [ 904.203884] env[61243]: _type = "Task" [ 904.203884] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.213081] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339063, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.236921] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2ca18506-4136-4cff-b747-8026c21178fa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.592446] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339060, 'name': PowerOnVM_Task, 'duration_secs': 0.519515} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.592823] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 904.593068] env[61243]: INFO nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Took 8.54 seconds to spawn the instance on the hypervisor. [ 904.593269] env[61243]: DEBUG nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.594346] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2799109f-cb88-4fdd-9ebe-a741fe57db75 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.614211] env[61243]: DEBUG nova.scheduler.client.report [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 105 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 904.614491] env[61243]: DEBUG nova.compute.provider_tree [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 105 to 106 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 904.614672] env[61243]: DEBUG nova.compute.provider_tree [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 904.631213] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339062, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.144895} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.632472] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.633342] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddec43d5-5637-46fe-8c87-f91b94b14a09 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.660417] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] d95a4d61-2bff-4d89-9ab8-f71e558f3409/d95a4d61-2bff-4d89-9ab8-f71e558f3409.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.661954] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c92f36e5-1906-4c3c-9729-af10b38d7ad4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.686360] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 904.686360] env[61243]: value = "task-1339064" [ 904.686360] env[61243]: _type = "Task" [ 904.686360] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.697896] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339064, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.716418] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339063, 'name': PowerOffVM_Task, 'duration_secs': 0.310355} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.717084] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 904.718106] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0916ae1-470e-43b7-b680-2e097c41dd2a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.741179] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f90bde63-f85d-4310-bc2b-221451379b52 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.114054] env[61243]: INFO nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Took 24.50 seconds to build instance. [ 905.119167] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.120039] env[61243]: DEBUG nova.compute.manager [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 905.122496] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.427s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.124024] env[61243]: INFO nova.compute.claims [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 905.197233] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339064, 'name': ReconfigVM_Task, 'duration_secs': 0.482659} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.197627] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Reconfigured VM instance instance-0000004f to attach disk [datastore2] d95a4d61-2bff-4d89-9ab8-f71e558f3409/d95a4d61-2bff-4d89-9ab8-f71e558f3409.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.198350] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-674a28ff-8a12-4ffe-b389-2ba1dadf0fc8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.206700] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 905.206700] env[61243]: value = "task-1339065" [ 905.206700] env[61243]: _type = "Task" [ 905.206700] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.216128] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339065, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.255804] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Creating Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 905.256325] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9b3f0186-77b5-43b5-969e-879ea0abf348 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.267283] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 905.267283] env[61243]: value = "task-1339066" [ 905.267283] env[61243]: _type = "Task" [ 905.267283] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.277864] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339066, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.616407] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "f8816b86-10c8-4d64-9659-fbc301893d46" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.013s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.631674] env[61243]: DEBUG nova.compute.utils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 905.635220] env[61243]: DEBUG nova.compute.manager [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 905.635398] env[61243]: DEBUG nova.network.neutron [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 905.688246] env[61243]: DEBUG nova.policy [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '785fd146024849efa0b67b1a7f731393', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9db958446334496bab832b95f2363f2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 905.719083] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339065, 'name': Rename_Task, 'duration_secs': 0.20991} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.720726] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 905.720726] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e7ebb202-4734-4383-a3b8-e7ec5e90836e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.727228] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 905.727228] env[61243]: value = "task-1339067" [ 905.727228] env[61243]: _type = "Task" [ 905.727228] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.736597] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339067, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.779955] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339066, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.969906] env[61243]: DEBUG nova.network.neutron [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Successfully created port: c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 906.134619] env[61243]: DEBUG nova.compute.manager [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 906.242233] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339067, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.283787] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339066, 'name': CreateSnapshot_Task, 'duration_secs': 1.012006} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.284102] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Created Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 906.284931] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe102285-2c7b-43a5-9757-3cdf54af5500 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.419054] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-640c76e1-3372-453e-a85c-d53214347738 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.427621] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dd6c630-6456-4f63-a3b2-a759ff722e07 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.461699] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac3a6ece-91f2-4ca0-95cd-4ab9a1c0761a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.471085] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb529d2a-c83e-4238-8384-108cdd0209e7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.487216] env[61243]: DEBUG nova.compute.provider_tree [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.739515] env[61243]: DEBUG oslo_vmware.api [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339067, 'name': PowerOnVM_Task, 'duration_secs': 0.649328} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.739826] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 906.740077] env[61243]: INFO nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Took 7.57 seconds to spawn the instance on the hypervisor. [ 906.740302] env[61243]: DEBUG nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.741164] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce63153-ef77-44bc-be95-f1b733d0c93c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.808540] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Creating linked-clone VM from snapshot {{(pid=61243) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 906.809376] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8404eb47-61d3-4524-bff5-2ed09b65dd32 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.819468] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 906.819468] env[61243]: value = "task-1339068" [ 906.819468] env[61243]: _type = "Task" [ 906.819468] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.829705] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339068, 'name': CloneVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.990821] env[61243]: DEBUG nova.scheduler.client.report [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.144868] env[61243]: DEBUG nova.compute.manager [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 907.260146] env[61243]: INFO nova.compute.manager [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Took 26.61 seconds to build instance. [ 907.332237] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339068, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.485434] env[61243]: DEBUG nova.compute.manager [req-901c99b1-4d26-409f-aabc-84463e6555e6 req-486189f1-f21e-41da-8bca-833b072d46c7 service nova] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Received event network-vif-plugged-c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 907.485673] env[61243]: DEBUG oslo_concurrency.lockutils [req-901c99b1-4d26-409f-aabc-84463e6555e6 req-486189f1-f21e-41da-8bca-833b072d46c7 service nova] Acquiring lock "e9a1c8af-7129-4366-bda1-1c3cf40eeab1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.485919] env[61243]: DEBUG oslo_concurrency.lockutils [req-901c99b1-4d26-409f-aabc-84463e6555e6 req-486189f1-f21e-41da-8bca-833b072d46c7 service nova] Lock "e9a1c8af-7129-4366-bda1-1c3cf40eeab1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.486120] env[61243]: DEBUG oslo_concurrency.lockutils [req-901c99b1-4d26-409f-aabc-84463e6555e6 req-486189f1-f21e-41da-8bca-833b072d46c7 service nova] Lock "e9a1c8af-7129-4366-bda1-1c3cf40eeab1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.486298] env[61243]: DEBUG nova.compute.manager [req-901c99b1-4d26-409f-aabc-84463e6555e6 req-486189f1-f21e-41da-8bca-833b072d46c7 service nova] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] No waiting events found dispatching network-vif-plugged-c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 907.486483] env[61243]: WARNING nova.compute.manager [req-901c99b1-4d26-409f-aabc-84463e6555e6 req-486189f1-f21e-41da-8bca-833b072d46c7 service nova] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Received unexpected event network-vif-plugged-c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3 for instance with vm_state building and task_state spawning. [ 907.496795] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.497304] env[61243]: DEBUG nova.compute.manager [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 907.500082] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.628s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.501694] env[61243]: INFO nova.compute.claims [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 907.541390] env[61243]: DEBUG nova.network.neutron [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Successfully updated port: c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 907.762119] env[61243]: DEBUG oslo_concurrency.lockutils [None req-47bedc53-dc9e-47a0-a260-74e67c41a573 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "d95a4d61-2bff-4d89-9ab8-f71e558f3409" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.120s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.831454] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339068, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.006589] env[61243]: DEBUG nova.compute.utils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 908.010894] env[61243]: DEBUG nova.compute.manager [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 908.011184] env[61243]: DEBUG nova.network.neutron [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 908.046146] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "refresh_cache-e9a1c8af-7129-4366-bda1-1c3cf40eeab1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.046146] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "refresh_cache-e9a1c8af-7129-4366-bda1-1c3cf40eeab1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.046146] env[61243]: DEBUG nova.network.neutron [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 908.063957] env[61243]: DEBUG nova.policy [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a9e6dc85a31d4ec38efe0c750c49ead1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '113713170fb94b8a8d7985b9a02faf77', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 908.333304] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339068, 'name': CloneVM_Task} progress is 95%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.435633] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "f8816b86-10c8-4d64-9659-fbc301893d46" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.436113] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "f8816b86-10c8-4d64-9659-fbc301893d46" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.436474] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "f8816b86-10c8-4d64-9659-fbc301893d46-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.436809] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "f8816b86-10c8-4d64-9659-fbc301893d46-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.437136] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "f8816b86-10c8-4d64-9659-fbc301893d46-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.440207] env[61243]: INFO nova.compute.manager [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Terminating instance [ 908.442730] env[61243]: DEBUG nova.compute.manager [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 908.443086] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 908.444429] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bdb5339-07ff-497c-981a-f643e8e23c13 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.456849] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 908.457258] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c215c8b-bc67-461f-89e5-8e10dfa4d702 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.465157] env[61243]: DEBUG oslo_vmware.api [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 908.465157] env[61243]: value = "task-1339069" [ 908.465157] env[61243]: _type = "Task" [ 908.465157] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.475101] env[61243]: DEBUG oslo_vmware.api [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339069, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.511631] env[61243]: DEBUG nova.compute.manager [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 908.601124] env[61243]: DEBUG nova.network.neutron [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 908.656028] env[61243]: DEBUG nova.network.neutron [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Successfully created port: f3206fdf-946a-4d55-b5a7-3b909da22db2 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 908.708382] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "d95a4d61-2bff-4d89-9ab8-f71e558f3409" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.708620] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "d95a4d61-2bff-4d89-9ab8-f71e558f3409" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.708895] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "d95a4d61-2bff-4d89-9ab8-f71e558f3409-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.709142] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "d95a4d61-2bff-4d89-9ab8-f71e558f3409-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.709397] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "d95a4d61-2bff-4d89-9ab8-f71e558f3409-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.715141] env[61243]: INFO nova.compute.manager [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Terminating instance [ 908.717695] env[61243]: DEBUG nova.compute.manager [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 908.718168] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 908.718984] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5188084-3e1c-47d0-8bee-77dba78b1f11 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.732258] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 908.732601] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49211bfa-aa20-4326-99bc-34955eaef169 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.740631] env[61243]: DEBUG oslo_vmware.api [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 908.740631] env[61243]: value = "task-1339070" [ 908.740631] env[61243]: _type = "Task" [ 908.740631] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.751328] env[61243]: DEBUG oslo_vmware.api [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339070, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.833073] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339068, 'name': CloneVM_Task, 'duration_secs': 1.733858} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.833390] env[61243]: INFO nova.virt.vmwareapi.vmops [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Created linked-clone VM from snapshot [ 908.834501] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81113285-91cb-4e7d-887d-54fa31922bc8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.843326] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Uploading image 0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 908.857797] env[61243]: DEBUG nova.network.neutron [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Updating instance_info_cache with network_info: [{"id": "c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3", "address": "fa:16:3e:c6:51:8f", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7f2cf3c-ba", "ovs_interfaceid": "c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.872496] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf68fd7-c88c-436b-89b5-2f2d90e1f3bf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.880553] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c346e71-0cf8-4b22-9155-7920b44b39c3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.913126] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666f3642-21a6-4674-83e0-997b06438e6e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.921943] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d11fe90-4eba-4e7f-824c-448846ee5831 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.937221] env[61243]: DEBUG nova.compute.provider_tree [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.976890] env[61243]: DEBUG oslo_vmware.api [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339069, 'name': PowerOffVM_Task, 'duration_secs': 0.23872} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.977357] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 908.977580] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 908.977868] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f7c38470-7447-4c2e-9879-8ef44aa8d917 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.044252] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 909.044503] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 909.044682] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Deleting the datastore file [datastore2] f8816b86-10c8-4d64-9659-fbc301893d46 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.044994] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ab64b31d-b7a3-473f-95af-728c786a0b62 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.051770] env[61243]: DEBUG oslo_vmware.api [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 909.051770] env[61243]: value = "task-1339072" [ 909.051770] env[61243]: _type = "Task" [ 909.051770] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.060687] env[61243]: DEBUG oslo_vmware.api [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339072, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.250879] env[61243]: DEBUG oslo_vmware.api [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339070, 'name': PowerOffVM_Task, 'duration_secs': 0.265857} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.251105] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 909.251296] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 909.251576] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7112fdda-d07d-4716-b572-c7a8a45bcb84 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.317612] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 909.317855] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 909.318080] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Deleting the datastore file [datastore2] d95a4d61-2bff-4d89-9ab8-f71e558f3409 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.318370] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3d409d23-ed1c-443a-bb72-6a9336ec115a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.324478] env[61243]: DEBUG oslo_vmware.api [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for the task: (returnval){ [ 909.324478] env[61243]: value = "task-1339074" [ 909.324478] env[61243]: _type = "Task" [ 909.324478] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.332917] env[61243]: DEBUG oslo_vmware.api [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339074, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.362712] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "refresh_cache-e9a1c8af-7129-4366-bda1-1c3cf40eeab1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.363109] env[61243]: DEBUG nova.compute.manager [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Instance network_info: |[{"id": "c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3", "address": "fa:16:3e:c6:51:8f", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7f2cf3c-ba", "ovs_interfaceid": "c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 909.440520] env[61243]: DEBUG nova.scheduler.client.report [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.523131] env[61243]: DEBUG nova.compute.manager [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 909.562051] env[61243]: DEBUG oslo_vmware.api [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339072, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152428} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.562352] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 909.562576] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 909.562771] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 909.562977] env[61243]: INFO nova.compute.manager [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Took 1.12 seconds to destroy the instance on the hypervisor. [ 909.563286] env[61243]: DEBUG oslo.service.loopingcall [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 909.563514] env[61243]: DEBUG nova.compute.manager [-] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 909.563616] env[61243]: DEBUG nova.network.neutron [-] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 909.717698] env[61243]: DEBUG nova.compute.manager [req-c787d43f-03b6-487b-b1ab-24ba23f1d000 req-eadbfb99-2d2d-4e90-a8ad-4ddfcaad4009 service nova] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Received event network-changed-c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.718163] env[61243]: DEBUG nova.compute.manager [req-c787d43f-03b6-487b-b1ab-24ba23f1d000 req-eadbfb99-2d2d-4e90-a8ad-4ddfcaad4009 service nova] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Refreshing instance network info cache due to event network-changed-c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 909.718488] env[61243]: DEBUG oslo_concurrency.lockutils [req-c787d43f-03b6-487b-b1ab-24ba23f1d000 req-eadbfb99-2d2d-4e90-a8ad-4ddfcaad4009 service nova] Acquiring lock "refresh_cache-e9a1c8af-7129-4366-bda1-1c3cf40eeab1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.718718] env[61243]: DEBUG oslo_concurrency.lockutils [req-c787d43f-03b6-487b-b1ab-24ba23f1d000 req-eadbfb99-2d2d-4e90-a8ad-4ddfcaad4009 service nova] Acquired lock "refresh_cache-e9a1c8af-7129-4366-bda1-1c3cf40eeab1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.719014] env[61243]: DEBUG nova.network.neutron [req-c787d43f-03b6-487b-b1ab-24ba23f1d000 req-eadbfb99-2d2d-4e90-a8ad-4ddfcaad4009 service nova] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Refreshing network info cache for port c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 909.837832] env[61243]: DEBUG oslo_vmware.api [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Task: {'id': task-1339074, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.182933} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.838146] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 909.838316] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 909.838548] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 909.838691] env[61243]: INFO nova.compute.manager [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Took 1.12 seconds to destroy the instance on the hypervisor. [ 909.838961] env[61243]: DEBUG oslo.service.loopingcall [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 909.839184] env[61243]: DEBUG nova.compute.manager [-] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 909.839279] env[61243]: DEBUG nova.network.neutron [-] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 909.945730] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.446s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.946285] env[61243]: DEBUG nova.compute.manager [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 909.949624] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.057s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.954476] env[61243]: INFO nova.compute.claims [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 910.460099] env[61243]: DEBUG nova.compute.utils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 910.464146] env[61243]: DEBUG nova.compute.manager [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 910.464237] env[61243]: DEBUG nova.network.neutron [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 910.490513] env[61243]: DEBUG nova.network.neutron [-] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.570368] env[61243]: DEBUG nova.policy [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32d9e3276f974c8f87573270909076ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7651b565d1647eb948fe07faba32a41', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 910.630469] env[61243]: DEBUG nova.network.neutron [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Successfully updated port: f3206fdf-946a-4d55-b5a7-3b909da22db2 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 910.712777] env[61243]: DEBUG nova.network.neutron [req-c787d43f-03b6-487b-b1ab-24ba23f1d000 req-eadbfb99-2d2d-4e90-a8ad-4ddfcaad4009 service nova] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Updated VIF entry in instance network info cache for port c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 910.713170] env[61243]: DEBUG nova.network.neutron [req-c787d43f-03b6-487b-b1ab-24ba23f1d000 req-eadbfb99-2d2d-4e90-a8ad-4ddfcaad4009 service nova] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Updating instance_info_cache with network_info: [{"id": "c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3", "address": "fa:16:3e:c6:51:8f", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7f2cf3c-ba", "ovs_interfaceid": "c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.787485] env[61243]: DEBUG nova.network.neutron [-] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.825163] env[61243]: DEBUG nova.virt.hardware [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.825648] env[61243]: DEBUG nova.virt.hardware [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.825881] env[61243]: DEBUG nova.virt.hardware [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.826150] env[61243]: DEBUG nova.virt.hardware [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.826346] env[61243]: DEBUG nova.virt.hardware [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.826509] env[61243]: DEBUG nova.virt.hardware [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.826978] env[61243]: DEBUG nova.virt.hardware [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.827188] env[61243]: DEBUG nova.virt.hardware [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.827405] env[61243]: DEBUG nova.virt.hardware [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.827608] env[61243]: DEBUG nova.virt.hardware [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.827796] env[61243]: DEBUG nova.virt.hardware [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.830266] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e3a3cb8-1396-4c36-8474-8d9d8e27382c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.839622] env[61243]: DEBUG oslo_vmware.rw_handles [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 910.839622] env[61243]: value = "vm-285734" [ 910.839622] env[61243]: _type = "VirtualMachine" [ 910.839622] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 910.840878] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a29ab7-b69f-44b4-a967-303e2a510efe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.845981] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-190136c5-f760-4a06-8b0a-0c3544d9cf54 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.858620] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:51:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a071ecf4-e713-4f97-9271-8c17952f6dee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 910.865982] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Creating folder: Project (c9db958446334496bab832b95f2363f2). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 910.869817] env[61243]: DEBUG nova.virt.hardware [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.869990] env[61243]: DEBUG nova.virt.hardware [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.870195] env[61243]: DEBUG nova.virt.hardware [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.870353] env[61243]: DEBUG nova.virt.hardware [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.870465] env[61243]: DEBUG nova.virt.hardware [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.870622] env[61243]: DEBUG nova.virt.hardware [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.870843] env[61243]: DEBUG nova.virt.hardware [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.871027] env[61243]: DEBUG nova.virt.hardware [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.871276] env[61243]: DEBUG nova.virt.hardware [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.871473] env[61243]: DEBUG nova.virt.hardware [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.871656] env[61243]: DEBUG nova.virt.hardware [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.875477] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b5ca56e3-d50b-4edb-9ada-31a79a4ec44a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.879449] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8922de-a411-4c9a-a34a-2824737d8ea1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.880113] env[61243]: DEBUG oslo_vmware.rw_handles [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lease: (returnval){ [ 910.880113] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522f01fa-ecf5-ca12-0d28-f7d8f2dc69c8" [ 910.880113] env[61243]: _type = "HttpNfcLease" [ 910.880113] env[61243]: } obtained for exporting VM: (result){ [ 910.880113] env[61243]: value = "vm-285734" [ 910.880113] env[61243]: _type = "VirtualMachine" [ 910.880113] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 910.880577] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the lease: (returnval){ [ 910.880577] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522f01fa-ecf5-ca12-0d28-f7d8f2dc69c8" [ 910.880577] env[61243]: _type = "HttpNfcLease" [ 910.880577] env[61243]: } to be ready. {{(pid=61243) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 910.887506] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1953ba1d-b544-424a-802b-5b4be7046cec {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.895777] env[61243]: DEBUG oslo_vmware.rw_handles [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524777f6-6cea-7f74-59c6-034ba8434556/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 910.895901] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 910.895901] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522f01fa-ecf5-ca12-0d28-f7d8f2dc69c8" [ 910.895901] env[61243]: _type = "HttpNfcLease" [ 910.895901] env[61243]: } is ready. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 910.896304] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Created folder: Project (c9db958446334496bab832b95f2363f2) in parent group-v285636. [ 910.896304] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Creating folder: Instances. Parent ref: group-v285735. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 910.897022] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c92965-96dc-4b44-a6f4-9e0b0cb4d833 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.900017] env[61243]: DEBUG oslo_vmware.rw_handles [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 910.900017] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522f01fa-ecf5-ca12-0d28-f7d8f2dc69c8" [ 910.900017] env[61243]: _type = "HttpNfcLease" [ 910.900017] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 910.900428] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dc9e279a-ef8a-4673-9cc2-4ddad24022b5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.910642] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4699958e-2d60-4125-ba9d-66d7ddf57b6c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.917794] env[61243]: DEBUG oslo_vmware.rw_handles [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524777f6-6cea-7f74-59c6-034ba8434556/disk-0.vmdk is in state: ready. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 910.917794] env[61243]: ERROR oslo_vmware.rw_handles [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524777f6-6cea-7f74-59c6-034ba8434556/disk-0.vmdk due to incomplete transfer. [ 910.919682] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-bf30cd87-dad5-4ab2-9e9f-970cc7dd3599 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.921273] env[61243]: DEBUG oslo_vmware.rw_handles [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f80849-2c16-7302-5e76-bc91067b1eeb/disk-0.vmdk from lease info. {{(pid=61243) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 910.921586] env[61243]: DEBUG oslo_vmware.rw_handles [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f80849-2c16-7302-5e76-bc91067b1eeb/disk-0.vmdk for reading. {{(pid=61243) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 910.924032] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Created folder: Instances in parent group-v285735. [ 910.924286] env[61243]: DEBUG oslo.service.loopingcall [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.925753] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 910.985442] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1e2bec88-feb2-46f3-aa84-178f96bae93c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.003619] env[61243]: DEBUG oslo_concurrency.lockutils [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "07ee984d-476e-484d-ba80-0ec2e411faa9" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.003962] env[61243]: DEBUG oslo_concurrency.lockutils [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.008048] env[61243]: DEBUG nova.compute.manager [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 911.014671] env[61243]: INFO nova.compute.manager [-] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Took 1.45 seconds to deallocate network for instance. [ 911.015328] env[61243]: DEBUG oslo_vmware.rw_handles [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524777f6-6cea-7f74-59c6-034ba8434556/disk-0.vmdk. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 911.015328] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Uploaded image 6fe66989-aad8-47fb-9027-b19aee66207e to the Glance image server {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 911.018067] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Destroying the VM {{(pid=61243) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 911.025209] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-74bcb1b9-5d7e-4e84-ab4c-58c6ef0d194e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.031991] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 911.031991] env[61243]: value = "task-1339079" [ 911.031991] env[61243]: _type = "Task" [ 911.031991] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.032253] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 911.032253] env[61243]: value = "task-1339078" [ 911.032253] env[61243]: _type = "Task" [ 911.032253] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.041599] env[61243]: DEBUG nova.network.neutron [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Successfully created port: bd92d46a-e97e-4169-a89e-59191cce73f3 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 911.052597] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339078, 'name': CreateVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.054215] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339079, 'name': Destroy_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.090327] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-16e96493-07ae-46b2-876f-98d26a7c4bb8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.141416] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "refresh_cache-ebe8cc8e-cc35-4ec4-bed7-d55df72a2991" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.141416] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "refresh_cache-ebe8cc8e-cc35-4ec4-bed7-d55df72a2991" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.141571] env[61243]: DEBUG nova.network.neutron [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 911.218596] env[61243]: DEBUG oslo_concurrency.lockutils [req-c787d43f-03b6-487b-b1ab-24ba23f1d000 req-eadbfb99-2d2d-4e90-a8ad-4ddfcaad4009 service nova] Releasing lock "refresh_cache-e9a1c8af-7129-4366-bda1-1c3cf40eeab1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.290464] env[61243]: INFO nova.compute.manager [-] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Took 1.45 seconds to deallocate network for instance. [ 911.390102] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dbf0944-7c26-4a74-9850-339aa6d8c028 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.403021] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ef68b9-e51c-49ff-aac1-208e47a27b96 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.438333] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c342d39-2ff7-4229-a02d-7d47d1f656ae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.445696] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8009d690-50a8-4963-bae2-047e233b3e86 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.461889] env[61243]: DEBUG nova.compute.provider_tree [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 911.512983] env[61243]: INFO nova.compute.manager [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Detaching volume 3111ce1e-7b01-4717-885c-55baf54556d1 [ 911.535682] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.566529] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339079, 'name': Destroy_Task, 'duration_secs': 0.449474} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.566529] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339078, 'name': CreateVM_Task, 'duration_secs': 0.410847} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.566756] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Destroyed the VM [ 911.566980] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Deleting Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 911.568790] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 911.568790] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5b5706ed-9f27-4295-bae9-bc7c5775d4f8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.572600] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.572834] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.574302] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 911.574302] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28c1981c-4881-4548-87a0-1b97bd5c6315 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.582152] env[61243]: INFO nova.virt.block_device [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Attempting to driver detach volume 3111ce1e-7b01-4717-885c-55baf54556d1 from mountpoint /dev/sdb [ 911.582502] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Volume detach. Driver type: vmdk {{(pid=61243) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 911.583038] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285711', 'volume_id': '3111ce1e-7b01-4717-885c-55baf54556d1', 'name': 'volume-3111ce1e-7b01-4717-885c-55baf54556d1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '07ee984d-476e-484d-ba80-0ec2e411faa9', 'attached_at': '', 'detached_at': '', 'volume_id': '3111ce1e-7b01-4717-885c-55baf54556d1', 'serial': '3111ce1e-7b01-4717-885c-55baf54556d1'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 911.587048] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b2f868-1b9c-450d-b95b-d5e3bf37a3b0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.598614] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 911.598614] env[61243]: value = "task-1339080" [ 911.598614] env[61243]: _type = "Task" [ 911.598614] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.606407] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 911.606407] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5241076a-fe2a-2df7-4190-062b8743ed93" [ 911.606407] env[61243]: _type = "Task" [ 911.606407] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.666539] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60905ce6-b0cc-4108-bfb9-0fa287e4b3f2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.685789] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339080, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.694159] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5241076a-fe2a-2df7-4190-062b8743ed93, 'name': SearchDatastore_Task, 'duration_secs': 0.012436} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.694159] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.694530] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 911.697750] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.697943] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.698188] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 911.702925] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a6ccfe48-ce31-4630-ae2e-4d6a74898956 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.705843] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f572ea4-6134-4634-a796-be44b42edf5b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.734986] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e672de4-446b-4e19-b8b9-03e42e45b31d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.738190] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 911.738190] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 911.738991] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ea0ab8b-e2f8-443d-9328-c94470a1ba05 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.754106] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] The volume has not been displaced from its original location: [datastore2] volume-3111ce1e-7b01-4717-885c-55baf54556d1/volume-3111ce1e-7b01-4717-885c-55baf54556d1.vmdk. No consolidation needed. {{(pid=61243) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 911.759652] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Reconfiguring VM instance instance-00000038 to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 911.761997] env[61243]: DEBUG nova.network.neutron [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 911.764301] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-698ca789-4c87-4bac-a431-c6ba1c103dd6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.777304] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 911.777304] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52861380-27e1-913c-0a54-91037bcf1ae0" [ 911.777304] env[61243]: _type = "Task" [ 911.777304] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.783088] env[61243]: DEBUG oslo_vmware.api [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 911.783088] env[61243]: value = "task-1339081" [ 911.783088] env[61243]: _type = "Task" [ 911.783088] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.792781] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52861380-27e1-913c-0a54-91037bcf1ae0, 'name': SearchDatastore_Task, 'duration_secs': 0.009545} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.794069] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1aa7d67-c3da-4ca2-9cf3-22be04021b5f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.800743] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.801060] env[61243]: DEBUG oslo_vmware.api [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339081, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.806700] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 911.806700] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e400a4-112b-8ae5-ea39-a37b58783397" [ 911.806700] env[61243]: _type = "Task" [ 911.806700] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.815772] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e400a4-112b-8ae5-ea39-a37b58783397, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.872573] env[61243]: DEBUG nova.compute.manager [req-14cf528d-03fd-4cd4-b0c7-b10ac26e0e23 req-ca4d61a4-de5f-43d1-8f85-c0f026c265e3 service nova] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Received event network-vif-deleted-e579287f-20a4-48c1-a93d-3b88d6da8af0 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.872866] env[61243]: DEBUG nova.compute.manager [req-14cf528d-03fd-4cd4-b0c7-b10ac26e0e23 req-ca4d61a4-de5f-43d1-8f85-c0f026c265e3 service nova] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Received event network-vif-deleted-f9010c84-231c-48c6-a2da-bcc9cd57ac3c {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.873083] env[61243]: DEBUG nova.compute.manager [req-14cf528d-03fd-4cd4-b0c7-b10ac26e0e23 req-ca4d61a4-de5f-43d1-8f85-c0f026c265e3 service nova] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Received event network-vif-plugged-f3206fdf-946a-4d55-b5a7-3b909da22db2 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.873386] env[61243]: DEBUG oslo_concurrency.lockutils [req-14cf528d-03fd-4cd4-b0c7-b10ac26e0e23 req-ca4d61a4-de5f-43d1-8f85-c0f026c265e3 service nova] Acquiring lock "ebe8cc8e-cc35-4ec4-bed7-d55df72a2991-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.873613] env[61243]: DEBUG oslo_concurrency.lockutils [req-14cf528d-03fd-4cd4-b0c7-b10ac26e0e23 req-ca4d61a4-de5f-43d1-8f85-c0f026c265e3 service nova] Lock "ebe8cc8e-cc35-4ec4-bed7-d55df72a2991-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.873865] env[61243]: DEBUG oslo_concurrency.lockutils [req-14cf528d-03fd-4cd4-b0c7-b10ac26e0e23 req-ca4d61a4-de5f-43d1-8f85-c0f026c265e3 service nova] Lock "ebe8cc8e-cc35-4ec4-bed7-d55df72a2991-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.874079] env[61243]: DEBUG nova.compute.manager [req-14cf528d-03fd-4cd4-b0c7-b10ac26e0e23 req-ca4d61a4-de5f-43d1-8f85-c0f026c265e3 service nova] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] No waiting events found dispatching network-vif-plugged-f3206fdf-946a-4d55-b5a7-3b909da22db2 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.874461] env[61243]: WARNING nova.compute.manager [req-14cf528d-03fd-4cd4-b0c7-b10ac26e0e23 req-ca4d61a4-de5f-43d1-8f85-c0f026c265e3 service nova] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Received unexpected event network-vif-plugged-f3206fdf-946a-4d55-b5a7-3b909da22db2 for instance with vm_state building and task_state spawning. [ 911.874709] env[61243]: DEBUG nova.compute.manager [req-14cf528d-03fd-4cd4-b0c7-b10ac26e0e23 req-ca4d61a4-de5f-43d1-8f85-c0f026c265e3 service nova] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Received event network-changed-f3206fdf-946a-4d55-b5a7-3b909da22db2 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.874901] env[61243]: DEBUG nova.compute.manager [req-14cf528d-03fd-4cd4-b0c7-b10ac26e0e23 req-ca4d61a4-de5f-43d1-8f85-c0f026c265e3 service nova] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Refreshing instance network info cache due to event network-changed-f3206fdf-946a-4d55-b5a7-3b909da22db2. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 911.875144] env[61243]: DEBUG oslo_concurrency.lockutils [req-14cf528d-03fd-4cd4-b0c7-b10ac26e0e23 req-ca4d61a4-de5f-43d1-8f85-c0f026c265e3 service nova] Acquiring lock "refresh_cache-ebe8cc8e-cc35-4ec4-bed7-d55df72a2991" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.997185] env[61243]: ERROR nova.scheduler.client.report [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [req-027150bb-1b7d-40d7-abd9-2f011f7bd208] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-027150bb-1b7d-40d7-abd9-2f011f7bd208"}]} [ 912.018017] env[61243]: DEBUG nova.network.neutron [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Updating instance_info_cache with network_info: [{"id": "f3206fdf-946a-4d55-b5a7-3b909da22db2", "address": "fa:16:3e:b7:ab:09", "network": {"id": "0788f9ac-5a15-4431-a711-9b4f6cf79fbc", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1210957657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "113713170fb94b8a8d7985b9a02faf77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c7821ea-f92f-4f06-a4cb-05e1186a9d22", "external-id": "nsx-vlan-transportzone-69", "segmentation_id": 69, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3206fdf-94", "ovs_interfaceid": "f3206fdf-946a-4d55-b5a7-3b909da22db2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.020328] env[61243]: DEBUG nova.scheduler.client.report [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 912.023941] env[61243]: DEBUG nova.compute.manager [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 912.045111] env[61243]: DEBUG nova.scheduler.client.report [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 912.045369] env[61243]: DEBUG nova.compute.provider_tree [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 912.065758] env[61243]: DEBUG nova.scheduler.client.report [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 912.092898] env[61243]: DEBUG nova.scheduler.client.report [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 912.115536] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339080, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.296030] env[61243]: DEBUG oslo_vmware.api [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339081, 'name': ReconfigVM_Task, 'duration_secs': 0.239237} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.296030] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Reconfigured VM instance instance-00000038 to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 912.300563] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d6799ff-2778-4310-b51d-df6637de0caa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.321368] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.321676] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.329137] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e400a4-112b-8ae5-ea39-a37b58783397, 'name': SearchDatastore_Task, 'duration_secs': 0.013741} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.332741] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.333082] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] e9a1c8af-7129-4366-bda1-1c3cf40eeab1/e9a1c8af-7129-4366-bda1-1c3cf40eeab1.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 912.333432] env[61243]: DEBUG oslo_vmware.api [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 912.333432] env[61243]: value = "task-1339082" [ 912.333432] env[61243]: _type = "Task" [ 912.333432] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.333814] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-272fb0b8-70ab-4b71-9dfa-c53e08900901 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.344992] env[61243]: DEBUG oslo_vmware.api [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339082, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.349454] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 912.349454] env[61243]: value = "task-1339083" [ 912.349454] env[61243]: _type = "Task" [ 912.349454] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.360766] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339083, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.406081] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e363c3-fca0-4b52-aebf-b1f1609976f5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.414309] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b60519-c476-4561-b23c-4e11cab9d0d5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.448995] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f0f0fe-a94e-4f3e-9e6f-2dfb035bc52d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.458011] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a507179a-099f-4795-b157-401868df6653 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.474496] env[61243]: DEBUG nova.compute.provider_tree [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 912.526486] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "refresh_cache-ebe8cc8e-cc35-4ec4-bed7-d55df72a2991" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.526869] env[61243]: DEBUG nova.compute.manager [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Instance network_info: |[{"id": "f3206fdf-946a-4d55-b5a7-3b909da22db2", "address": "fa:16:3e:b7:ab:09", "network": {"id": "0788f9ac-5a15-4431-a711-9b4f6cf79fbc", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1210957657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "113713170fb94b8a8d7985b9a02faf77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c7821ea-f92f-4f06-a4cb-05e1186a9d22", "external-id": "nsx-vlan-transportzone-69", "segmentation_id": 69, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3206fdf-94", "ovs_interfaceid": "f3206fdf-946a-4d55-b5a7-3b909da22db2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 912.527248] env[61243]: DEBUG oslo_concurrency.lockutils [req-14cf528d-03fd-4cd4-b0c7-b10ac26e0e23 req-ca4d61a4-de5f-43d1-8f85-c0f026c265e3 service nova] Acquired lock "refresh_cache-ebe8cc8e-cc35-4ec4-bed7-d55df72a2991" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.527544] env[61243]: DEBUG nova.network.neutron [req-14cf528d-03fd-4cd4-b0c7-b10ac26e0e23 req-ca4d61a4-de5f-43d1-8f85-c0f026c265e3 service nova] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Refreshing network info cache for port f3206fdf-946a-4d55-b5a7-3b909da22db2 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 912.528870] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:ab:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5c7821ea-f92f-4f06-a4cb-05e1186a9d22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f3206fdf-946a-4d55-b5a7-3b909da22db2', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.536514] env[61243]: DEBUG oslo.service.loopingcall [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.540518] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 912.541889] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e9f829ad-32e8-487d-b9be-49218c062c81 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.563919] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 912.563919] env[61243]: value = "task-1339084" [ 912.563919] env[61243]: _type = "Task" [ 912.563919] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.573420] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339084, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.623920] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339080, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.826311] env[61243]: DEBUG nova.virt.hardware [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.830276] env[61243]: DEBUG nova.virt.hardware [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.830276] env[61243]: DEBUG nova.virt.hardware [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.830276] env[61243]: DEBUG nova.virt.hardware [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.830276] env[61243]: DEBUG nova.virt.hardware [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.830276] env[61243]: DEBUG nova.virt.hardware [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.830276] env[61243]: DEBUG nova.virt.hardware [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.830276] env[61243]: DEBUG nova.virt.hardware [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.830276] env[61243]: DEBUG nova.virt.hardware [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.830276] env[61243]: DEBUG nova.virt.hardware [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.830276] env[61243]: DEBUG nova.virt.hardware [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.830276] env[61243]: DEBUG nova.compute.manager [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 912.838314] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-309acc48-b1bb-4d9d-9e2c-ceccd8a3bfad {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.865214] env[61243]: DEBUG oslo_vmware.rw_handles [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52fd5ec2-92bb-9b45-a78f-c49232bad080/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 912.871311] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292427df-71ce-423f-b97f-abb490ef3d3e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.875796] env[61243]: DEBUG oslo_vmware.api [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339082, 'name': ReconfigVM_Task, 'duration_secs': 0.14489} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.878660] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-303fb0d1-8491-43c2-802c-dfde206f73ae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.884898] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285711', 'volume_id': '3111ce1e-7b01-4717-885c-55baf54556d1', 'name': 'volume-3111ce1e-7b01-4717-885c-55baf54556d1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '07ee984d-476e-484d-ba80-0ec2e411faa9', 'attached_at': '', 'detached_at': '', 'volume_id': '3111ce1e-7b01-4717-885c-55baf54556d1', 'serial': '3111ce1e-7b01-4717-885c-55baf54556d1'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 912.893178] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339083, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.894973] env[61243]: DEBUG oslo_vmware.rw_handles [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52fd5ec2-92bb-9b45-a78f-c49232bad080/disk-0.vmdk is in state: ready. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 912.895229] env[61243]: ERROR oslo_vmware.rw_handles [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52fd5ec2-92bb-9b45-a78f-c49232bad080/disk-0.vmdk due to incomplete transfer. [ 912.906021] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-4a4dfe4a-e201-40e9-936d-a7b50881e088 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.913297] env[61243]: DEBUG oslo_vmware.rw_handles [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52fd5ec2-92bb-9b45-a78f-c49232bad080/disk-0.vmdk. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 912.913297] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Uploaded image 9a127f45-caa9-4072-8a3a-d6944f9047b8 to the Glance image server {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 912.914875] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Destroying the VM {{(pid=61243) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 912.915256] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ee5d190b-506b-4dab-bb42-fbecc50c8019 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.925146] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 912.925146] env[61243]: value = "task-1339085" [ 912.925146] env[61243]: _type = "Task" [ 912.925146] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.936771] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339085, 'name': Destroy_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.989113] env[61243]: DEBUG nova.network.neutron [req-14cf528d-03fd-4cd4-b0c7-b10ac26e0e23 req-ca4d61a4-de5f-43d1-8f85-c0f026c265e3 service nova] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Updated VIF entry in instance network info cache for port f3206fdf-946a-4d55-b5a7-3b909da22db2. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 912.989501] env[61243]: DEBUG nova.network.neutron [req-14cf528d-03fd-4cd4-b0c7-b10ac26e0e23 req-ca4d61a4-de5f-43d1-8f85-c0f026c265e3 service nova] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Updating instance_info_cache with network_info: [{"id": "f3206fdf-946a-4d55-b5a7-3b909da22db2", "address": "fa:16:3e:b7:ab:09", "network": {"id": "0788f9ac-5a15-4431-a711-9b4f6cf79fbc", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1210957657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "113713170fb94b8a8d7985b9a02faf77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c7821ea-f92f-4f06-a4cb-05e1186a9d22", "external-id": "nsx-vlan-transportzone-69", "segmentation_id": 69, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3206fdf-94", "ovs_interfaceid": "f3206fdf-946a-4d55-b5a7-3b909da22db2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.006398] env[61243]: DEBUG nova.network.neutron [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Successfully updated port: bd92d46a-e97e-4169-a89e-59191cce73f3 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 913.028905] env[61243]: DEBUG nova.scheduler.client.report [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 107 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 913.030423] env[61243]: DEBUG nova.compute.provider_tree [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 107 to 108 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 913.030423] env[61243]: DEBUG nova.compute.provider_tree [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 913.075189] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339084, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.116051] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339080, 'name': RemoveSnapshot_Task} progress is 36%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.361961] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339083, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.562253} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.362366] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] e9a1c8af-7129-4366-bda1-1c3cf40eeab1/e9a1c8af-7129-4366-bda1-1c3cf40eeab1.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 913.362511] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 913.362889] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-170be042-96f1-4e1f-951a-99a89b91da17 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.369494] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 913.369494] env[61243]: value = "task-1339086" [ 913.369494] env[61243]: _type = "Task" [ 913.369494] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.380020] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339086, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.380020] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.435120] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339085, 'name': Destroy_Task, 'duration_secs': 0.3476} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.435400] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Destroyed the VM [ 913.435636] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Deleting Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 913.435896] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-bcb2f8a2-d592-4ce7-8f9c-7c7dd1854ab4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.443279] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 913.443279] env[61243]: value = "task-1339087" [ 913.443279] env[61243]: _type = "Task" [ 913.443279] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.451605] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339087, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.472012] env[61243]: DEBUG nova.objects.instance [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lazy-loading 'flavor' on Instance uuid 07ee984d-476e-484d-ba80-0ec2e411faa9 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.491940] env[61243]: DEBUG oslo_concurrency.lockutils [req-14cf528d-03fd-4cd4-b0c7-b10ac26e0e23 req-ca4d61a4-de5f-43d1-8f85-c0f026c265e3 service nova] Releasing lock "refresh_cache-ebe8cc8e-cc35-4ec4-bed7-d55df72a2991" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.509740] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "refresh_cache-d46fb6dc-cdd9-4688-b5e7-9efde0f770c9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.509869] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquired lock "refresh_cache-d46fb6dc-cdd9-4688-b5e7-9efde0f770c9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.510094] env[61243]: DEBUG nova.network.neutron [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 913.535869] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.586s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.535869] env[61243]: DEBUG nova.compute.manager [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 913.538614] env[61243]: DEBUG oslo_concurrency.lockutils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.573s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.540514] env[61243]: INFO nova.compute.claims [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 913.575877] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339084, 'name': CreateVM_Task, 'duration_secs': 0.886536} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.576166] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 913.577089] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.577574] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.577721] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 913.578030] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96e562db-c305-4cf8-9dc5-465352a98f97 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.584157] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 913.584157] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e98c36-05d0-a21f-3c4c-06d7a9a5b196" [ 913.584157] env[61243]: _type = "Task" [ 913.584157] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.592198] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e98c36-05d0-a21f-3c4c-06d7a9a5b196, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.616250] env[61243]: DEBUG oslo_vmware.api [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339080, 'name': RemoveSnapshot_Task, 'duration_secs': 1.684352} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.616567] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Deleted Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 913.616799] env[61243]: INFO nova.compute.manager [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Took 14.95 seconds to snapshot the instance on the hypervisor. [ 913.880511] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339086, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.179549} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.881266] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 913.881927] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1024c81e-b242-4b4a-be4a-3c80f6338c4a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.904181] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] e9a1c8af-7129-4366-bda1-1c3cf40eeab1/e9a1c8af-7129-4366-bda1-1c3cf40eeab1.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.904314] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ecd851b0-0645-4784-9828-8624f4804c11 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.922859] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 913.922859] env[61243]: value = "task-1339088" [ 913.922859] env[61243]: _type = "Task" [ 913.922859] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.931195] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339088, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.952856] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339087, 'name': RemoveSnapshot_Task, 'duration_secs': 0.37371} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.953165] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Deleted Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 913.953452] env[61243]: DEBUG nova.compute.manager [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 913.954251] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8dfd342-a20b-462c-a17b-f611332f4567 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.983093] env[61243]: DEBUG nova.compute.manager [req-b1488135-e7d2-4ee3-ac11-19c2b46d2268 req-97fc6f7b-3c3c-401a-95f5-552a317d87db service nova] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Received event network-vif-plugged-bd92d46a-e97e-4169-a89e-59191cce73f3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.983093] env[61243]: DEBUG oslo_concurrency.lockutils [req-b1488135-e7d2-4ee3-ac11-19c2b46d2268 req-97fc6f7b-3c3c-401a-95f5-552a317d87db service nova] Acquiring lock "d46fb6dc-cdd9-4688-b5e7-9efde0f770c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.983093] env[61243]: DEBUG oslo_concurrency.lockutils [req-b1488135-e7d2-4ee3-ac11-19c2b46d2268 req-97fc6f7b-3c3c-401a-95f5-552a317d87db service nova] Lock "d46fb6dc-cdd9-4688-b5e7-9efde0f770c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.983093] env[61243]: DEBUG oslo_concurrency.lockutils [req-b1488135-e7d2-4ee3-ac11-19c2b46d2268 req-97fc6f7b-3c3c-401a-95f5-552a317d87db service nova] Lock "d46fb6dc-cdd9-4688-b5e7-9efde0f770c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.983093] env[61243]: DEBUG nova.compute.manager [req-b1488135-e7d2-4ee3-ac11-19c2b46d2268 req-97fc6f7b-3c3c-401a-95f5-552a317d87db service nova] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] No waiting events found dispatching network-vif-plugged-bd92d46a-e97e-4169-a89e-59191cce73f3 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 913.983093] env[61243]: WARNING nova.compute.manager [req-b1488135-e7d2-4ee3-ac11-19c2b46d2268 req-97fc6f7b-3c3c-401a-95f5-552a317d87db service nova] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Received unexpected event network-vif-plugged-bd92d46a-e97e-4169-a89e-59191cce73f3 for instance with vm_state building and task_state spawning. [ 913.983093] env[61243]: DEBUG nova.compute.manager [req-b1488135-e7d2-4ee3-ac11-19c2b46d2268 req-97fc6f7b-3c3c-401a-95f5-552a317d87db service nova] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Received event network-changed-bd92d46a-e97e-4169-a89e-59191cce73f3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.983093] env[61243]: DEBUG nova.compute.manager [req-b1488135-e7d2-4ee3-ac11-19c2b46d2268 req-97fc6f7b-3c3c-401a-95f5-552a317d87db service nova] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Refreshing instance network info cache due to event network-changed-bd92d46a-e97e-4169-a89e-59191cce73f3. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 913.983093] env[61243]: DEBUG oslo_concurrency.lockutils [req-b1488135-e7d2-4ee3-ac11-19c2b46d2268 req-97fc6f7b-3c3c-401a-95f5-552a317d87db service nova] Acquiring lock "refresh_cache-d46fb6dc-cdd9-4688-b5e7-9efde0f770c9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.048223] env[61243]: DEBUG nova.compute.utils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 914.050469] env[61243]: DEBUG nova.network.neutron [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 914.052733] env[61243]: DEBUG nova.compute.manager [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 914.052905] env[61243]: DEBUG nova.network.neutron [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 914.094022] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e98c36-05d0-a21f-3c4c-06d7a9a5b196, 'name': SearchDatastore_Task, 'duration_secs': 0.016536} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.094160] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.094735] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.094735] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.094923] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.094984] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.095271] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0e163c76-8344-49f2-b5fd-bb0edd50d39c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.105774] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.106014] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 914.107039] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c292617-0260-46e1-b049-71858785d3f6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.113587] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 914.113587] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525dcc18-a76c-b72e-cff2-4546520053bf" [ 914.113587] env[61243]: _type = "Task" [ 914.113587] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.126646] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525dcc18-a76c-b72e-cff2-4546520053bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.134198] env[61243]: DEBUG nova.policy [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06e247d66c8d43218302a35ead439405', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e641e72e4d2048218ef403456dbb2b82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 914.225757] env[61243]: DEBUG nova.compute.manager [None req-840a8f7b-a1f8-4a7e-b037-148a0fd1f567 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Found 1 images (rotation: 2) {{(pid=61243) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 914.240535] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f0aa87f4-414f-4a27-aa15-43ac93b0b1d0 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "07ee984d-476e-484d-ba80-0ec2e411faa9" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.274895] env[61243]: DEBUG nova.network.neutron [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Updating instance_info_cache with network_info: [{"id": "bd92d46a-e97e-4169-a89e-59191cce73f3", "address": "fa:16:3e:97:d0:81", "network": {"id": "523d4cf1-467c-4ea6-80a4-ebbd01016efa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-919325376-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7651b565d1647eb948fe07faba32a41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd92d46a-e9", "ovs_interfaceid": "bd92d46a-e97e-4169-a89e-59191cce73f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.434683] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339088, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.468636] env[61243]: INFO nova.compute.manager [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Shelve offloading [ 914.471520] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 914.471520] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-de182878-c7d8-4a06-8435-e003990a814d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.478348] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 914.478348] env[61243]: value = "task-1339089" [ 914.478348] env[61243]: _type = "Task" [ 914.478348] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.485278] env[61243]: DEBUG nova.network.neutron [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Successfully created port: 133a7a1c-7381-4bac-8300-4bee50500e62 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 914.487360] env[61243]: DEBUG oslo_concurrency.lockutils [None req-96a4d2df-9203-4bdc-bae4-48f7d02b2e98 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.483s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.489777] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f0aa87f4-414f-4a27-aa15-43ac93b0b1d0 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.250s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.490044] env[61243]: DEBUG nova.compute.manager [None req-f0aa87f4-414f-4a27-aa15-43ac93b0b1d0 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.491757] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3626cdb-9b3f-40b2-8bb7-bc3791e38edd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.499434] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] VM already powered off {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 914.499729] env[61243]: DEBUG nova.compute.manager [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.500564] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9fb971-f2c0-4391-b0d2-558060d637f6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.511086] env[61243]: DEBUG nova.compute.manager [None req-f0aa87f4-414f-4a27-aa15-43ac93b0b1d0 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61243) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 914.511979] env[61243]: DEBUG nova.objects.instance [None req-f0aa87f4-414f-4a27-aa15-43ac93b0b1d0 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lazy-loading 'flavor' on Instance uuid 07ee984d-476e-484d-ba80-0ec2e411faa9 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.521212] env[61243]: DEBUG oslo_concurrency.lockutils [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.521212] env[61243]: DEBUG oslo_concurrency.lockutils [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.521212] env[61243]: DEBUG nova.network.neutron [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 914.555396] env[61243]: DEBUG nova.compute.manager [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 914.626272] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525dcc18-a76c-b72e-cff2-4546520053bf, 'name': SearchDatastore_Task, 'duration_secs': 0.015842} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.627116] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95f3d9c6-75de-45c9-acb8-06feeb548635 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.638566] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 914.638566] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525d7c12-c27d-fef4-37ec-df8fa0ba0c86" [ 914.638566] env[61243]: _type = "Task" [ 914.638566] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.644343] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525d7c12-c27d-fef4-37ec-df8fa0ba0c86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.777527] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Releasing lock "refresh_cache-d46fb6dc-cdd9-4688-b5e7-9efde0f770c9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.777779] env[61243]: DEBUG nova.compute.manager [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Instance network_info: |[{"id": "bd92d46a-e97e-4169-a89e-59191cce73f3", "address": "fa:16:3e:97:d0:81", "network": {"id": "523d4cf1-467c-4ea6-80a4-ebbd01016efa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-919325376-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7651b565d1647eb948fe07faba32a41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd92d46a-e9", "ovs_interfaceid": "bd92d46a-e97e-4169-a89e-59191cce73f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 914.778117] env[61243]: DEBUG oslo_concurrency.lockutils [req-b1488135-e7d2-4ee3-ac11-19c2b46d2268 req-97fc6f7b-3c3c-401a-95f5-552a317d87db service nova] Acquired lock "refresh_cache-d46fb6dc-cdd9-4688-b5e7-9efde0f770c9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.778370] env[61243]: DEBUG nova.network.neutron [req-b1488135-e7d2-4ee3-ac11-19c2b46d2268 req-97fc6f7b-3c3c-401a-95f5-552a317d87db service nova] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Refreshing network info cache for port bd92d46a-e97e-4169-a89e-59191cce73f3 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 914.779631] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:97:d0:81', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '42f08482-a1da-405d-9918-d733d9f5173c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd92d46a-e97e-4169-a89e-59191cce73f3', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.788223] env[61243]: DEBUG oslo.service.loopingcall [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.793698] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 914.794470] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8bf2d020-59df-48eb-825b-6394fc01b1ae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.817965] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.817965] env[61243]: value = "task-1339090" [ 914.817965] env[61243]: _type = "Task" [ 914.817965] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.828038] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339090, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.847702] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ff67908-2d94-4626-8dad-6d8f07f93716 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.854790] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e42cb5a-a2d5-427d-8256-03cae42a3098 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.891030] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0834f360-a560-416b-9025-c3c9da419e6a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.899325] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa25b3c4-f6a7-4a61-b51b-1c6d0cf53ea0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.914795] env[61243]: DEBUG nova.compute.provider_tree [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 914.935776] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339088, 'name': ReconfigVM_Task, 'duration_secs': 0.619348} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.936094] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Reconfigured VM instance instance-00000050 to attach disk [datastore2] e9a1c8af-7129-4366-bda1-1c3cf40eeab1/e9a1c8af-7129-4366-bda1-1c3cf40eeab1.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.936866] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57fe0837-12e8-41fe-83fd-d24cf5534f00 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.943307] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 914.943307] env[61243]: value = "task-1339091" [ 914.943307] env[61243]: _type = "Task" [ 914.943307] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.953816] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339091, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.016935] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0aa87f4-414f-4a27-aa15-43ac93b0b1d0 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 915.017242] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ccebc019-fa23-404b-808c-9d1691ff0d14 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.026708] env[61243]: DEBUG oslo_vmware.api [None req-f0aa87f4-414f-4a27-aa15-43ac93b0b1d0 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 915.026708] env[61243]: value = "task-1339092" [ 915.026708] env[61243]: _type = "Task" [ 915.026708] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.037180] env[61243]: DEBUG oslo_vmware.api [None req-f0aa87f4-414f-4a27-aa15-43ac93b0b1d0 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339092, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.080955] env[61243]: DEBUG nova.network.neutron [req-b1488135-e7d2-4ee3-ac11-19c2b46d2268 req-97fc6f7b-3c3c-401a-95f5-552a317d87db service nova] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Updated VIF entry in instance network info cache for port bd92d46a-e97e-4169-a89e-59191cce73f3. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 915.081407] env[61243]: DEBUG nova.network.neutron [req-b1488135-e7d2-4ee3-ac11-19c2b46d2268 req-97fc6f7b-3c3c-401a-95f5-552a317d87db service nova] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Updating instance_info_cache with network_info: [{"id": "bd92d46a-e97e-4169-a89e-59191cce73f3", "address": "fa:16:3e:97:d0:81", "network": {"id": "523d4cf1-467c-4ea6-80a4-ebbd01016efa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-919325376-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7651b565d1647eb948fe07faba32a41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd92d46a-e9", "ovs_interfaceid": "bd92d46a-e97e-4169-a89e-59191cce73f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.149976] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525d7c12-c27d-fef4-37ec-df8fa0ba0c86, 'name': SearchDatastore_Task, 'duration_secs': 0.043262} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.150368] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.150694] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] ebe8cc8e-cc35-4ec4-bed7-d55df72a2991/ebe8cc8e-cc35-4ec4-bed7-d55df72a2991.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 915.151028] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-59755a89-0b55-4cad-87d1-64365eba608a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.167294] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 915.167294] env[61243]: value = "task-1339093" [ 915.167294] env[61243]: _type = "Task" [ 915.167294] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.175936] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339093, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.318960] env[61243]: DEBUG nova.network.neutron [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Updating instance_info_cache with network_info: [{"id": "970e8509-164d-4ddf-8f0c-795e92883f6b", "address": "fa:16:3e:00:af:7d", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap970e8509-16", "ovs_interfaceid": "970e8509-164d-4ddf-8f0c-795e92883f6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.332231] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339090, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.453775] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339091, 'name': Rename_Task, 'duration_secs': 0.246343} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.454011] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 915.454289] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5660b9c4-a0d7-4ce0-8e13-e18d03f711ca {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.461027] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 915.461027] env[61243]: value = "task-1339094" [ 915.461027] env[61243]: _type = "Task" [ 915.461027] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.465011] env[61243]: DEBUG nova.scheduler.client.report [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 108 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 915.465295] env[61243]: DEBUG nova.compute.provider_tree [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 108 to 109 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 915.465502] env[61243]: DEBUG nova.compute.provider_tree [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 915.476397] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339094, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.538366] env[61243]: DEBUG oslo_vmware.api [None req-f0aa87f4-414f-4a27-aa15-43ac93b0b1d0 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339092, 'name': PowerOffVM_Task, 'duration_secs': 0.396467} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.538855] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0aa87f4-414f-4a27-aa15-43ac93b0b1d0 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 915.539230] env[61243]: DEBUG nova.compute.manager [None req-f0aa87f4-414f-4a27-aa15-43ac93b0b1d0 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.540360] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de9e1d5-daf2-4bf5-84e8-ab87a0fa85e4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.568373] env[61243]: DEBUG nova.compute.manager [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 915.584279] env[61243]: DEBUG oslo_concurrency.lockutils [req-b1488135-e7d2-4ee3-ac11-19c2b46d2268 req-97fc6f7b-3c3c-401a-95f5-552a317d87db service nova] Releasing lock "refresh_cache-d46fb6dc-cdd9-4688-b5e7-9efde0f770c9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.594206] env[61243]: DEBUG nova.virt.hardware [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 915.594494] env[61243]: DEBUG nova.virt.hardware [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 915.594651] env[61243]: DEBUG nova.virt.hardware [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 915.594846] env[61243]: DEBUG nova.virt.hardware [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 915.594995] env[61243]: DEBUG nova.virt.hardware [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 915.595193] env[61243]: DEBUG nova.virt.hardware [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 915.595421] env[61243]: DEBUG nova.virt.hardware [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 915.595606] env[61243]: DEBUG nova.virt.hardware [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 915.595784] env[61243]: DEBUG nova.virt.hardware [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 915.595956] env[61243]: DEBUG nova.virt.hardware [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 915.596159] env[61243]: DEBUG nova.virt.hardware [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 915.597093] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-281d3fdd-354c-47c8-bcaa-24e0dfdb59e1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.606853] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a34a68c-2cb5-45d9-84a4-6a904bbc2605 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.678557] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339093, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.786626] env[61243]: DEBUG nova.compute.manager [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.786626] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd44b0a-cf67-4e51-97b5-f75f82cae776 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.827515] env[61243]: DEBUG oslo_concurrency.lockutils [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Releasing lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.835144] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339090, 'name': CreateVM_Task, 'duration_secs': 0.571792} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.835740] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 915.836500] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.836680] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.837045] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.837665] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bac6ca69-9f88-4747-af04-4af255cb83ac {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.845065] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 915.845065] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c1757a-408d-4e51-11f9-0593ca3827bc" [ 915.845065] env[61243]: _type = "Task" [ 915.845065] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.856826] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c1757a-408d-4e51-11f9-0593ca3827bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.942882] env[61243]: DEBUG nova.compute.manager [req-13ca3224-2887-43a1-8d09-55af35d16d43 req-b0c6f91a-964e-4b5f-ad38-8193c71c7f80 service nova] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Received event network-vif-plugged-133a7a1c-7381-4bac-8300-4bee50500e62 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.942882] env[61243]: DEBUG oslo_concurrency.lockutils [req-13ca3224-2887-43a1-8d09-55af35d16d43 req-b0c6f91a-964e-4b5f-ad38-8193c71c7f80 service nova] Acquiring lock "5c063b4a-3fbb-4222-ab6d-7d27b6b99084-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.942882] env[61243]: DEBUG oslo_concurrency.lockutils [req-13ca3224-2887-43a1-8d09-55af35d16d43 req-b0c6f91a-964e-4b5f-ad38-8193c71c7f80 service nova] Lock "5c063b4a-3fbb-4222-ab6d-7d27b6b99084-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.942882] env[61243]: DEBUG oslo_concurrency.lockutils [req-13ca3224-2887-43a1-8d09-55af35d16d43 req-b0c6f91a-964e-4b5f-ad38-8193c71c7f80 service nova] Lock "5c063b4a-3fbb-4222-ab6d-7d27b6b99084-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.944257] env[61243]: DEBUG nova.compute.manager [req-13ca3224-2887-43a1-8d09-55af35d16d43 req-b0c6f91a-964e-4b5f-ad38-8193c71c7f80 service nova] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] No waiting events found dispatching network-vif-plugged-133a7a1c-7381-4bac-8300-4bee50500e62 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 915.944257] env[61243]: WARNING nova.compute.manager [req-13ca3224-2887-43a1-8d09-55af35d16d43 req-b0c6f91a-964e-4b5f-ad38-8193c71c7f80 service nova] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Received unexpected event network-vif-plugged-133a7a1c-7381-4bac-8300-4bee50500e62 for instance with vm_state building and task_state spawning. [ 915.973086] env[61243]: DEBUG oslo_concurrency.lockutils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.973503] env[61243]: DEBUG nova.compute.manager [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 915.976376] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339094, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.979176] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.419s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.979176] env[61243]: DEBUG nova.objects.instance [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Lazy-loading 'resources' on Instance uuid 36db1ed5-846f-4ad6-8cee-38b73ff00321 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.988637] env[61243]: DEBUG nova.network.neutron [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Successfully updated port: 133a7a1c-7381-4bac-8300-4bee50500e62 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 916.057632] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f0aa87f4-414f-4a27-aa15-43ac93b0b1d0 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.565s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.062298] env[61243]: DEBUG nova.compute.manager [req-077a199f-0d75-47a4-9372-c0cf4af57340 req-b1b80ab6-6af8-4e13-91b0-1b54ff022ebb service nova] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Received event network-changed-133a7a1c-7381-4bac-8300-4bee50500e62 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.062298] env[61243]: DEBUG nova.compute.manager [req-077a199f-0d75-47a4-9372-c0cf4af57340 req-b1b80ab6-6af8-4e13-91b0-1b54ff022ebb service nova] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Refreshing instance network info cache due to event network-changed-133a7a1c-7381-4bac-8300-4bee50500e62. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 916.062298] env[61243]: DEBUG oslo_concurrency.lockutils [req-077a199f-0d75-47a4-9372-c0cf4af57340 req-b1b80ab6-6af8-4e13-91b0-1b54ff022ebb service nova] Acquiring lock "refresh_cache-5c063b4a-3fbb-4222-ab6d-7d27b6b99084" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.062298] env[61243]: DEBUG oslo_concurrency.lockutils [req-077a199f-0d75-47a4-9372-c0cf4af57340 req-b1b80ab6-6af8-4e13-91b0-1b54ff022ebb service nova] Acquired lock "refresh_cache-5c063b4a-3fbb-4222-ab6d-7d27b6b99084" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.062298] env[61243]: DEBUG nova.network.neutron [req-077a199f-0d75-47a4-9372-c0cf4af57340 req-b1b80ab6-6af8-4e13-91b0-1b54ff022ebb service nova] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Refreshing network info cache for port 133a7a1c-7381-4bac-8300-4bee50500e62 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 916.182287] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339093, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.716309} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.183152] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] ebe8cc8e-cc35-4ec4-bed7-d55df72a2991/ebe8cc8e-cc35-4ec4-bed7-d55df72a2991.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 916.183492] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 916.183669] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd08d44b-f381-43ab-8eec-266ec1df0ac0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.191260] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 916.191260] env[61243]: value = "task-1339095" [ 916.191260] env[61243]: _type = "Task" [ 916.191260] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.200492] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339095, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.301270] env[61243]: INFO nova.compute.manager [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] instance snapshotting [ 916.302013] env[61243]: DEBUG nova.objects.instance [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lazy-loading 'flavor' on Instance uuid 316572e0-c007-42cb-aaf0-3a8cfcaf24aa {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.312775] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 916.313719] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0967fe89-70f4-4d89-826a-81c8577d06c7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.322695] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 916.322975] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2c940bd-efa2-4f5d-a979-5e0900422040 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.358048] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c1757a-408d-4e51-11f9-0593ca3827bc, 'name': SearchDatastore_Task, 'duration_secs': 0.020652} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.358581] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.358978] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 916.359442] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.359674] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.359944] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 916.360743] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-22d4f03b-c79f-405b-96d1-99a09768689e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.373633] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 916.373849] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 916.374663] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a24e9ac3-7b9a-49a1-a974-03c658465bde {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.380245] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 916.380245] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e39f25-6608-cb0c-f4e4-148015fc5f16" [ 916.380245] env[61243]: _type = "Task" [ 916.380245] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.389984] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e39f25-6608-cb0c-f4e4-148015fc5f16, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.420112] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 916.420112] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 916.420368] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Deleting the datastore file [datastore1] c1a0d242-4eae-4c03-8341-840b41341f17 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.420496] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-069dec92-e175-4fe9-84e0-c1e9cd989946 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.427048] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 916.427048] env[61243]: value = "task-1339097" [ 916.427048] env[61243]: _type = "Task" [ 916.427048] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.434703] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339097, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.471902] env[61243]: DEBUG oslo_vmware.api [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339094, 'name': PowerOnVM_Task, 'duration_secs': 0.969844} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.472330] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 916.472686] env[61243]: INFO nova.compute.manager [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Took 9.33 seconds to spawn the instance on the hypervisor. [ 916.472926] env[61243]: DEBUG nova.compute.manager [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 916.473752] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc08062-4b89-43c8-a154-b6d1a9c2d121 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.478913] env[61243]: DEBUG nova.compute.utils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 916.481806] env[61243]: DEBUG nova.compute.manager [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 916.482077] env[61243]: DEBUG nova.network.neutron [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 916.494033] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "refresh_cache-5c063b4a-3fbb-4222-ab6d-7d27b6b99084" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.533291] env[61243]: DEBUG nova.policy [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cc77f359ba44dc0bc8e053d5a4b63ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3248cca8a8aa4cad84b8bd5dcf7ed0c3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 916.606808] env[61243]: DEBUG nova.network.neutron [req-077a199f-0d75-47a4-9372-c0cf4af57340 req-b1b80ab6-6af8-4e13-91b0-1b54ff022ebb service nova] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 916.672386] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Acquiring lock "8df486c4-d1ce-47b0-81a8-ad9e83eaac7f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.672632] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Lock "8df486c4-d1ce-47b0-81a8-ad9e83eaac7f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.672913] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Acquiring lock "8df486c4-d1ce-47b0-81a8-ad9e83eaac7f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.673263] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Lock "8df486c4-d1ce-47b0-81a8-ad9e83eaac7f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.673487] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Lock "8df486c4-d1ce-47b0-81a8-ad9e83eaac7f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.675796] env[61243]: INFO nova.compute.manager [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Terminating instance [ 916.679663] env[61243]: DEBUG nova.compute.manager [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 916.679881] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 916.681169] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04fd1fa-746d-48ea-8db6-44eec1b2f3d7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.692577] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 916.692905] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-45b4f513-2c80-4add-a7f0-9500411fc021 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.702754] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339095, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.131142} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.704104] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 916.704487] env[61243]: DEBUG oslo_vmware.api [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Waiting for the task: (returnval){ [ 916.704487] env[61243]: value = "task-1339098" [ 916.704487] env[61243]: _type = "Task" [ 916.704487] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.708964] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b6aeb3f-42a1-4a1d-9141-056805a8c62b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.719706] env[61243]: DEBUG oslo_vmware.api [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1339098, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.737299] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] ebe8cc8e-cc35-4ec4-bed7-d55df72a2991/ebe8cc8e-cc35-4ec4-bed7-d55df72a2991.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.740022] env[61243]: DEBUG nova.objects.instance [None req-cf852253-6dcc-4dd2-853a-894a99c83073 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lazy-loading 'flavor' on Instance uuid 07ee984d-476e-484d-ba80-0ec2e411faa9 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.741426] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da7c05b0-4bdb-4874-96f6-8b8666efd406 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.762695] env[61243]: DEBUG nova.network.neutron [req-077a199f-0d75-47a4-9372-c0cf4af57340 req-b1b80ab6-6af8-4e13-91b0-1b54ff022ebb service nova] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.767602] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cf852253-6dcc-4dd2-853a-894a99c83073 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.767759] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cf852253-6dcc-4dd2-853a-894a99c83073 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquired lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.767915] env[61243]: DEBUG nova.network.neutron [None req-cf852253-6dcc-4dd2-853a-894a99c83073 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 916.768117] env[61243]: DEBUG nova.objects.instance [None req-cf852253-6dcc-4dd2-853a-894a99c83073 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lazy-loading 'info_cache' on Instance uuid 07ee984d-476e-484d-ba80-0ec2e411faa9 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.776063] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 916.776063] env[61243]: value = "task-1339099" [ 916.776063] env[61243]: _type = "Task" [ 916.776063] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.789924] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339099, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.807296] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07a7f89-8096-4956-ba54-f6ad5321cf73 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.811726] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01b43f7-1287-42bc-af26-66d3bca4107a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.837680] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb4aa34-f948-4d0f-bd2c-400da0118109 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.842939] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79ff4956-e943-4fb5-926e-800d8ae3755f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.882046] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-877ef698-e504-4d7e-8434-0e6f22fdf826 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.888526] env[61243]: DEBUG nova.network.neutron [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Successfully created port: 99fff1a2-f410-40f3-9e32-fa6bb307a34a {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.894707] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaec364e-9520-4509-82a6-d65ec23d0d30 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.904247] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e39f25-6608-cb0c-f4e4-148015fc5f16, 'name': SearchDatastore_Task, 'duration_secs': 0.017673} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.905983] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d5c33d8-cc0b-4ecc-9706-5857e68dfe80 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.919382] env[61243]: DEBUG nova.compute.provider_tree [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.925041] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 916.925041] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52aaf587-7a62-2696-6a14-2b30df87c21c" [ 916.925041] env[61243]: _type = "Task" [ 916.925041] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.940029] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52aaf587-7a62-2696-6a14-2b30df87c21c, 'name': SearchDatastore_Task, 'duration_secs': 0.011134} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.941990] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.942320] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] d46fb6dc-cdd9-4688-b5e7-9efde0f770c9/d46fb6dc-cdd9-4688-b5e7-9efde0f770c9.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 916.942633] env[61243]: DEBUG oslo_vmware.api [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339097, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.317326} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.943173] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-00f57ba6-7822-4127-b868-ce0854c675d6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.945706] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 916.945912] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 916.946126] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 916.953881] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 916.953881] env[61243]: value = "task-1339100" [ 916.953881] env[61243]: _type = "Task" [ 916.953881] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.963450] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339100, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.972874] env[61243]: INFO nova.scheduler.client.report [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Deleted allocations for instance c1a0d242-4eae-4c03-8341-840b41341f17 [ 916.990684] env[61243]: DEBUG nova.compute.manager [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 917.001161] env[61243]: INFO nova.compute.manager [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Took 28.70 seconds to build instance. [ 917.224678] env[61243]: DEBUG oslo_vmware.api [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1339098, 'name': PowerOffVM_Task, 'duration_secs': 0.258398} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.225009] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 917.225198] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 917.225479] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-401a1c00-fa6e-4ecd-a007-640e6c94aaad {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.268459] env[61243]: DEBUG oslo_concurrency.lockutils [req-077a199f-0d75-47a4-9372-c0cf4af57340 req-b1b80ab6-6af8-4e13-91b0-1b54ff022ebb service nova] Releasing lock "refresh_cache-5c063b4a-3fbb-4222-ab6d-7d27b6b99084" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.268836] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "refresh_cache-5c063b4a-3fbb-4222-ab6d-7d27b6b99084" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.268999] env[61243]: DEBUG nova.network.neutron [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 917.271894] env[61243]: DEBUG nova.objects.base [None req-cf852253-6dcc-4dd2-853a-894a99c83073 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Object Instance<07ee984d-476e-484d-ba80-0ec2e411faa9> lazy-loaded attributes: flavor,info_cache {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 917.286974] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339099, 'name': ReconfigVM_Task, 'duration_secs': 0.384023} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.287482] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Reconfigured VM instance instance-00000051 to attach disk [datastore2] ebe8cc8e-cc35-4ec4-bed7-d55df72a2991/ebe8cc8e-cc35-4ec4-bed7-d55df72a2991.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.288448] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-77aa7979-0af7-40bb-9da6-9c1a1242ed75 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.296506] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 917.296506] env[61243]: value = "task-1339102" [ 917.296506] env[61243]: _type = "Task" [ 917.296506] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.301627] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 917.301861] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 917.302062] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Deleting the datastore file [datastore2] 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 917.302742] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ab120ce-9d5b-46fd-ac68-4baf384794d8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.308511] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339102, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.313910] env[61243]: DEBUG oslo_vmware.api [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Waiting for the task: (returnval){ [ 917.313910] env[61243]: value = "task-1339103" [ 917.313910] env[61243]: _type = "Task" [ 917.313910] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.323544] env[61243]: DEBUG oslo_vmware.api [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1339103, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.382389] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Creating Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 917.382701] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c0e0be57-25fc-4f9d-829c-eb0557371752 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.393864] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 917.393864] env[61243]: value = "task-1339104" [ 917.393864] env[61243]: _type = "Task" [ 917.393864] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.402984] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339104, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.423924] env[61243]: DEBUG nova.scheduler.client.report [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.466128] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339100, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.477784] env[61243]: DEBUG oslo_concurrency.lockutils [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.507027] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bcce5c6e-628f-4167-9295-6ade1b716133 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "e9a1c8af-7129-4366-bda1-1c3cf40eeab1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.209s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.806794] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339102, 'name': Rename_Task, 'duration_secs': 0.339652} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.807070] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 917.807177] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d7095a14-ade1-47d1-a85c-ed93de547e75 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.813362] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 917.813362] env[61243]: value = "task-1339105" [ 917.813362] env[61243]: _type = "Task" [ 917.813362] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.820379] env[61243]: DEBUG nova.network.neutron [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 917.827790] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339105, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.831752] env[61243]: DEBUG oslo_vmware.api [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Task: {'id': task-1339103, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.370922} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.832015] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.832231] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 917.832411] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 917.832585] env[61243]: INFO nova.compute.manager [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 917.832831] env[61243]: DEBUG oslo.service.loopingcall [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.833030] env[61243]: DEBUG nova.compute.manager [-] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 917.833141] env[61243]: DEBUG nova.network.neutron [-] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 917.907543] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339104, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.933888] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.956s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.939501] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.114s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.939501] env[61243]: DEBUG nova.objects.instance [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lazy-loading 'resources' on Instance uuid 6036199d-cda9-42f8-8669-bc099d2b5ffe {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.966251] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339100, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.598405} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.966613] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] d46fb6dc-cdd9-4688-b5e7-9efde0f770c9/d46fb6dc-cdd9-4688-b5e7-9efde0f770c9.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 917.966968] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 917.967341] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e092a08f-ae6d-4b63-88f5-caa2cc63f915 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.976287] env[61243]: INFO nova.scheduler.client.report [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Deleted allocations for instance 36db1ed5-846f-4ad6-8cee-38b73ff00321 [ 917.988531] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 917.988531] env[61243]: value = "task-1339106" [ 917.988531] env[61243]: _type = "Task" [ 917.988531] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.001944] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339106, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.005483] env[61243]: DEBUG nova.compute.manager [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 918.045203] env[61243]: DEBUG nova.virt.hardware [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='b83f1b0cd4b3c2c587f3b4f3ec3e509e',container_format='bare',created_at=2024-10-22T18:09:52Z,direct_url=,disk_format='vmdk',id=2b9c3b0a-5f89-462e-99b2-0c88500ad15b,min_disk=1,min_ram=0,name='tempest-test-snap-1160391834',owner='3248cca8a8aa4cad84b8bd5dcf7ed0c3',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-22T18:10:07Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 918.046771] env[61243]: DEBUG nova.virt.hardware [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 918.046771] env[61243]: DEBUG nova.virt.hardware [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 918.046771] env[61243]: DEBUG nova.virt.hardware [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 918.046771] env[61243]: DEBUG nova.virt.hardware [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 918.046771] env[61243]: DEBUG nova.virt.hardware [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 918.046771] env[61243]: DEBUG nova.virt.hardware [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 918.047057] env[61243]: DEBUG nova.virt.hardware [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 918.047133] env[61243]: DEBUG nova.virt.hardware [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 918.047305] env[61243]: DEBUG nova.virt.hardware [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 918.047478] env[61243]: DEBUG nova.virt.hardware [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 918.048769] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c90e219a-3da8-4b15-9a10-8f41af7a047e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.061593] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-528dcb58-5354-4702-9bc4-53510ed64906 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.195546] env[61243]: DEBUG nova.network.neutron [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Updating instance_info_cache with network_info: [{"id": "133a7a1c-7381-4bac-8300-4bee50500e62", "address": "fa:16:3e:a9:6e:a3", "network": {"id": "bce1aecb-e043-4f68-99e2-c6b03c4bb866", "bridge": "br-int", "label": "tempest-ServersTestJSON-652904947-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e641e72e4d2048218ef403456dbb2b82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap133a7a1c-73", "ovs_interfaceid": "133a7a1c-7381-4bac-8300-4bee50500e62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.325374] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339105, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.392229] env[61243]: DEBUG nova.compute.manager [req-2f957cc1-303c-4eca-9be6-9d6b11bf45c1 req-24499ee0-4b16-497f-a416-fcd42c56ddb8 service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Received event network-vif-unplugged-970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.392229] env[61243]: DEBUG oslo_concurrency.lockutils [req-2f957cc1-303c-4eca-9be6-9d6b11bf45c1 req-24499ee0-4b16-497f-a416-fcd42c56ddb8 service nova] Acquiring lock "c1a0d242-4eae-4c03-8341-840b41341f17-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.392229] env[61243]: DEBUG oslo_concurrency.lockutils [req-2f957cc1-303c-4eca-9be6-9d6b11bf45c1 req-24499ee0-4b16-497f-a416-fcd42c56ddb8 service nova] Lock "c1a0d242-4eae-4c03-8341-840b41341f17-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.392229] env[61243]: DEBUG oslo_concurrency.lockutils [req-2f957cc1-303c-4eca-9be6-9d6b11bf45c1 req-24499ee0-4b16-497f-a416-fcd42c56ddb8 service nova] Lock "c1a0d242-4eae-4c03-8341-840b41341f17-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.392229] env[61243]: DEBUG nova.compute.manager [req-2f957cc1-303c-4eca-9be6-9d6b11bf45c1 req-24499ee0-4b16-497f-a416-fcd42c56ddb8 service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] No waiting events found dispatching network-vif-unplugged-970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.392865] env[61243]: WARNING nova.compute.manager [req-2f957cc1-303c-4eca-9be6-9d6b11bf45c1 req-24499ee0-4b16-497f-a416-fcd42c56ddb8 service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Received unexpected event network-vif-unplugged-970e8509-164d-4ddf-8f0c-795e92883f6b for instance with vm_state shelved_offloaded and task_state None. [ 918.393425] env[61243]: DEBUG nova.compute.manager [req-2f957cc1-303c-4eca-9be6-9d6b11bf45c1 req-24499ee0-4b16-497f-a416-fcd42c56ddb8 service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Received event network-changed-970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.393975] env[61243]: DEBUG nova.compute.manager [req-2f957cc1-303c-4eca-9be6-9d6b11bf45c1 req-24499ee0-4b16-497f-a416-fcd42c56ddb8 service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Refreshing instance network info cache due to event network-changed-970e8509-164d-4ddf-8f0c-795e92883f6b. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 918.395963] env[61243]: DEBUG oslo_concurrency.lockutils [req-2f957cc1-303c-4eca-9be6-9d6b11bf45c1 req-24499ee0-4b16-497f-a416-fcd42c56ddb8 service nova] Acquiring lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.395963] env[61243]: DEBUG oslo_concurrency.lockutils [req-2f957cc1-303c-4eca-9be6-9d6b11bf45c1 req-24499ee0-4b16-497f-a416-fcd42c56ddb8 service nova] Acquired lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.395963] env[61243]: DEBUG nova.network.neutron [req-2f957cc1-303c-4eca-9be6-9d6b11bf45c1 req-24499ee0-4b16-497f-a416-fcd42c56ddb8 service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Refreshing network info cache for port 970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 918.415240] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339104, 'name': CreateSnapshot_Task, 'duration_secs': 0.878767} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.416553] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Created Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 918.417470] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e54e80f-27d1-4648-8f4b-0f0b2a841e3f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.479786] env[61243]: INFO nova.compute.manager [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Rebuilding instance [ 918.489692] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3aeb62e7-bb53-40e9-965f-6f141547eea6 tempest-AttachInterfacesUnderV243Test-1713149283 tempest-AttachInterfacesUnderV243Test-1713149283-project-member] Lock "36db1ed5-846f-4ad6-8cee-38b73ff00321" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.302s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.503637] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339106, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.126623} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.504031] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 918.504998] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a60687-27bc-49fd-b293-a86aeecff311 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.537491] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] d46fb6dc-cdd9-4688-b5e7-9efde0f770c9/d46fb6dc-cdd9-4688-b5e7-9efde0f770c9.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 918.546248] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e644871e-4287-4c85-b2fa-f3588b9a7fea {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.569659] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 918.569659] env[61243]: value = "task-1339107" [ 918.569659] env[61243]: _type = "Task" [ 918.569659] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.574714] env[61243]: DEBUG nova.network.neutron [None req-cf852253-6dcc-4dd2-853a-894a99c83073 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Updating instance_info_cache with network_info: [{"id": "af7195a1-9a59-4b3a-b62d-7e907c94ce7b", "address": "fa:16:3e:ee:b7:e1", "network": {"id": "c6d4943d-4a8d-41b4-8023-c4f1549d312d", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1045259606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8549626a07e941e1ab03452355a0f116", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf7195a1-9a", "ovs_interfaceid": "af7195a1-9a59-4b3a-b62d-7e907c94ce7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.591668] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339107, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.592905] env[61243]: DEBUG nova.compute.manager [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.593805] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14019d92-6d08-462b-bf99-126336773443 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.700350] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "refresh_cache-5c063b4a-3fbb-4222-ab6d-7d27b6b99084" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.700350] env[61243]: DEBUG nova.compute.manager [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Instance network_info: |[{"id": "133a7a1c-7381-4bac-8300-4bee50500e62", "address": "fa:16:3e:a9:6e:a3", "network": {"id": "bce1aecb-e043-4f68-99e2-c6b03c4bb866", "bridge": "br-int", "label": "tempest-ServersTestJSON-652904947-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e641e72e4d2048218ef403456dbb2b82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap133a7a1c-73", "ovs_interfaceid": "133a7a1c-7381-4bac-8300-4bee50500e62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 918.700691] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:6e:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '90878b7b-ddb7-4f47-892b-d6e06f73475f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '133a7a1c-7381-4bac-8300-4bee50500e62', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 918.709892] env[61243]: DEBUG oslo.service.loopingcall [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.710152] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 918.712988] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4e0cf766-41a9-42ee-b6e6-14f35d6cac97 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.739848] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 918.739848] env[61243]: value = "task-1339108" [ 918.739848] env[61243]: _type = "Task" [ 918.739848] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.753648] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339108, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.823546] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d25fab7-42bb-453b-8402-55473abb2678 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.833416] env[61243]: DEBUG oslo_vmware.api [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339105, 'name': PowerOnVM_Task, 'duration_secs': 0.590376} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.835764] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 918.836035] env[61243]: INFO nova.compute.manager [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Took 9.31 seconds to spawn the instance on the hypervisor. [ 918.836250] env[61243]: DEBUG nova.compute.manager [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.838026] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b1f61f-22da-4235-8542-6c4ccdac240a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.842526] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe76cec5-86f4-46e0-9f9b-1972e0d185c2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.884661] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6a354eb-7817-4e00-b0e8-57bd2f2ecd3a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.890527] env[61243]: DEBUG oslo_concurrency.lockutils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "f77c2610-fa49-487c-9fb4-3b112f800daf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.890814] env[61243]: DEBUG oslo_concurrency.lockutils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "f77c2610-fa49-487c-9fb4-3b112f800daf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.899305] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f82d91d3-031a-42fd-8b23-0710791e5e43 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.918949] env[61243]: DEBUG nova.compute.provider_tree [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.937777] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Creating linked-clone VM from snapshot {{(pid=61243) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 918.938480] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e9717157-8b99-423a-92fc-b2cd83bfe412 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.947987] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 918.947987] env[61243]: value = "task-1339109" [ 918.947987] env[61243]: _type = "Task" [ 918.947987] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.960060] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339109, 'name': CloneVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.986558] env[61243]: DEBUG nova.network.neutron [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Successfully updated port: 99fff1a2-f410-40f3-9e32-fa6bb307a34a {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 919.077884] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cf852253-6dcc-4dd2-853a-894a99c83073 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Releasing lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.089603] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339107, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.109974] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 919.111053] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-36cc3d13-5e88-4a5f-8c5b-00160e7a81c3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.122643] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 919.122643] env[61243]: value = "task-1339110" [ 919.122643] env[61243]: _type = "Task" [ 919.122643] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.137683] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339110, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.249945] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339108, 'name': CreateVM_Task, 'duration_secs': 0.434393} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.252842] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 919.253115] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.253436] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.253769] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 919.254394] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8088ef3-4150-4626-af2f-d8188e4e6c83 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.259664] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 919.259664] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e0d962-1257-7bcf-2d2b-a1e73fd067c0" [ 919.259664] env[61243]: _type = "Task" [ 919.259664] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.269050] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e0d962-1257-7bcf-2d2b-a1e73fd067c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.293451] env[61243]: DEBUG nova.network.neutron [req-2f957cc1-303c-4eca-9be6-9d6b11bf45c1 req-24499ee0-4b16-497f-a416-fcd42c56ddb8 service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Updated VIF entry in instance network info cache for port 970e8509-164d-4ddf-8f0c-795e92883f6b. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 919.293649] env[61243]: DEBUG nova.network.neutron [req-2f957cc1-303c-4eca-9be6-9d6b11bf45c1 req-24499ee0-4b16-497f-a416-fcd42c56ddb8 service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Updating instance_info_cache with network_info: [{"id": "970e8509-164d-4ddf-8f0c-795e92883f6b", "address": "fa:16:3e:00:af:7d", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap970e8509-16", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.395985] env[61243]: DEBUG nova.compute.manager [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 919.405181] env[61243]: INFO nova.compute.manager [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Took 25.73 seconds to build instance. [ 919.421859] env[61243]: DEBUG nova.scheduler.client.report [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.458442] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339109, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.490267] env[61243]: DEBUG oslo_concurrency.lockutils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "refresh_cache-09587d68-f7fd-4648-bd5c-e53b1b6464e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.490463] env[61243]: DEBUG oslo_concurrency.lockutils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired lock "refresh_cache-09587d68-f7fd-4648-bd5c-e53b1b6464e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.490623] env[61243]: DEBUG nova.network.neutron [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 919.557850] env[61243]: DEBUG nova.network.neutron [-] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.584548] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339107, 'name': ReconfigVM_Task, 'duration_secs': 0.553959} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.584548] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Reconfigured VM instance instance-00000052 to attach disk [datastore2] d46fb6dc-cdd9-4688-b5e7-9efde0f770c9/d46fb6dc-cdd9-4688-b5e7-9efde0f770c9.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 919.584548] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-44ade56c-7cb8-48da-a0ed-1553be72899e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.589623] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 919.589623] env[61243]: value = "task-1339111" [ 919.589623] env[61243]: _type = "Task" [ 919.589623] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.595028] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf852253-6dcc-4dd2-853a-894a99c83073 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 919.595028] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-daa16c99-4bb4-40a2-81d7-a6cfa62a8080 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.604955] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339111, 'name': Rename_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.606522] env[61243]: DEBUG oslo_vmware.api [None req-cf852253-6dcc-4dd2-853a-894a99c83073 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 919.606522] env[61243]: value = "task-1339112" [ 919.606522] env[61243]: _type = "Task" [ 919.606522] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.616471] env[61243]: DEBUG oslo_vmware.api [None req-cf852253-6dcc-4dd2-853a-894a99c83073 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339112, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.636796] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339110, 'name': PowerOffVM_Task, 'duration_secs': 0.21266} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.636951] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 919.637267] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 919.638437] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79958d59-cc6c-432c-8595-8a69f7896ce2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.649592] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 919.649963] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-facadc01-f4e8-49eb-9e40-ea06d2f1e932 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.735233] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 919.735686] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 919.736140] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleting the datastore file [datastore2] e9a1c8af-7129-4366-bda1-1c3cf40eeab1 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 919.736569] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e60e01a7-7d15-4102-bb01-95b7efd8d6d1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.750038] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 919.750038] env[61243]: value = "task-1339114" [ 919.750038] env[61243]: _type = "Task" [ 919.750038] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.758959] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339114, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.778126] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e0d962-1257-7bcf-2d2b-a1e73fd067c0, 'name': SearchDatastore_Task, 'duration_secs': 0.016494} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.780692] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.781186] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.781550] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.781830] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.782211] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.782673] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fbcb709c-1ac8-424a-b8a3-68ee2cd0dad9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.797251] env[61243]: DEBUG oslo_concurrency.lockutils [req-2f957cc1-303c-4eca-9be6-9d6b11bf45c1 req-24499ee0-4b16-497f-a416-fcd42c56ddb8 service nova] Releasing lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.799017] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.799370] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 919.800496] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c6784cd-7f45-48dd-8988-e6d169a2e9e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.809687] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 919.809687] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523f6dd1-9f4a-926e-9e99-4ef75a28b52b" [ 919.809687] env[61243]: _type = "Task" [ 919.809687] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.819210] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523f6dd1-9f4a-926e-9e99-4ef75a28b52b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.912582] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d3ad0f7-174b-4fbd-bf6b-e873cc6cc4df tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "ebe8cc8e-cc35-4ec4-bed7-d55df72a2991" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.245s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.928035] env[61243]: DEBUG oslo_concurrency.lockutils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.928859] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.990s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.931551] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.397s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.931828] env[61243]: DEBUG nova.objects.instance [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lazy-loading 'resources' on Instance uuid f8816b86-10c8-4d64-9659-fbc301893d46 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.949504] env[61243]: INFO nova.scheduler.client.report [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Deleted allocations for instance 6036199d-cda9-42f8-8669-bc099d2b5ffe [ 919.967287] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339109, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.042209] env[61243]: DEBUG nova.network.neutron [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 920.063604] env[61243]: INFO nova.compute.manager [-] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Took 2.23 seconds to deallocate network for instance. [ 920.100404] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339111, 'name': Rename_Task, 'duration_secs': 0.22346} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.100759] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 920.101080] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aeab79ba-0c98-4724-b66d-dc9a78448ae6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.115418] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 920.115418] env[61243]: value = "task-1339115" [ 920.115418] env[61243]: _type = "Task" [ 920.115418] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.124143] env[61243]: DEBUG oslo_vmware.api [None req-cf852253-6dcc-4dd2-853a-894a99c83073 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339112, 'name': PowerOnVM_Task, 'duration_secs': 0.474529} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.124874] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf852253-6dcc-4dd2-853a-894a99c83073 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 920.125212] env[61243]: DEBUG nova.compute.manager [None req-cf852253-6dcc-4dd2-853a-894a99c83073 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.126065] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee18093-4995-4119-bf29-724f85fb814f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.132358] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339115, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.262352] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339114, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.319554} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.262638] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 920.262840] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 920.263061] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 920.284514] env[61243]: DEBUG nova.network.neutron [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Updating instance_info_cache with network_info: [{"id": "99fff1a2-f410-40f3-9e32-fa6bb307a34a", "address": "fa:16:3e:44:ae:bb", "network": {"id": "2737f3f1-a8da-4b5f-a33d-c768f14207e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1004846100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3248cca8a8aa4cad84b8bd5dcf7ed0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99fff1a2-f4", "ovs_interfaceid": "99fff1a2-f410-40f3-9e32-fa6bb307a34a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.322569] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523f6dd1-9f4a-926e-9e99-4ef75a28b52b, 'name': SearchDatastore_Task, 'duration_secs': 0.012879} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.323497] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49780a58-bdcc-4c56-b353-3e0ffc73bedc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.328844] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 920.328844] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e0c642-eed3-b8c2-7815-80d3bd3763af" [ 920.328844] env[61243]: _type = "Task" [ 920.328844] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.337054] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e0c642-eed3-b8c2-7815-80d3bd3763af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.442191] env[61243]: DEBUG oslo_concurrency.lockutils [None req-093ad0fe-8eb7-42a3-9348-e903beb275aa tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "ebe8cc8e-cc35-4ec4-bed7-d55df72a2991" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.442476] env[61243]: DEBUG oslo_concurrency.lockutils [None req-093ad0fe-8eb7-42a3-9348-e903beb275aa tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "ebe8cc8e-cc35-4ec4-bed7-d55df72a2991" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.442685] env[61243]: DEBUG nova.compute.manager [None req-093ad0fe-8eb7-42a3-9348-e903beb275aa tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.443649] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2470bf2-db68-450e-973e-6deee99cf04c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.450046] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "c1a0d242-4eae-4c03-8341-840b41341f17" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.456952] env[61243]: DEBUG nova.compute.manager [None req-093ad0fe-8eb7-42a3-9348-e903beb275aa tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61243) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 920.457525] env[61243]: DEBUG nova.objects.instance [None req-093ad0fe-8eb7-42a3-9348-e903beb275aa tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lazy-loading 'flavor' on Instance uuid ebe8cc8e-cc35-4ec4-bed7-d55df72a2991 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 920.465787] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339109, 'name': CloneVM_Task} progress is 95%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.468825] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b7962854-dbc2-49e5-af61-bab3083adbb2 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "6036199d-cda9-42f8-8669-bc099d2b5ffe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.636s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.508537] env[61243]: DEBUG nova.compute.manager [req-ea11618d-616e-4111-822a-bfee05fe412a req-8d21566f-63a5-49e6-9a2b-944055efbc7b service nova] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Received event network-vif-plugged-99fff1a2-f410-40f3-9e32-fa6bb307a34a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.508946] env[61243]: DEBUG oslo_concurrency.lockutils [req-ea11618d-616e-4111-822a-bfee05fe412a req-8d21566f-63a5-49e6-9a2b-944055efbc7b service nova] Acquiring lock "09587d68-f7fd-4648-bd5c-e53b1b6464e8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.509187] env[61243]: DEBUG oslo_concurrency.lockutils [req-ea11618d-616e-4111-822a-bfee05fe412a req-8d21566f-63a5-49e6-9a2b-944055efbc7b service nova] Lock "09587d68-f7fd-4648-bd5c-e53b1b6464e8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.509418] env[61243]: DEBUG oslo_concurrency.lockutils [req-ea11618d-616e-4111-822a-bfee05fe412a req-8d21566f-63a5-49e6-9a2b-944055efbc7b service nova] Lock "09587d68-f7fd-4648-bd5c-e53b1b6464e8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.509672] env[61243]: DEBUG nova.compute.manager [req-ea11618d-616e-4111-822a-bfee05fe412a req-8d21566f-63a5-49e6-9a2b-944055efbc7b service nova] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] No waiting events found dispatching network-vif-plugged-99fff1a2-f410-40f3-9e32-fa6bb307a34a {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 920.509789] env[61243]: WARNING nova.compute.manager [req-ea11618d-616e-4111-822a-bfee05fe412a req-8d21566f-63a5-49e6-9a2b-944055efbc7b service nova] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Received unexpected event network-vif-plugged-99fff1a2-f410-40f3-9e32-fa6bb307a34a for instance with vm_state building and task_state spawning. [ 920.509975] env[61243]: DEBUG nova.compute.manager [req-ea11618d-616e-4111-822a-bfee05fe412a req-8d21566f-63a5-49e6-9a2b-944055efbc7b service nova] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Received event network-changed-99fff1a2-f410-40f3-9e32-fa6bb307a34a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.510169] env[61243]: DEBUG nova.compute.manager [req-ea11618d-616e-4111-822a-bfee05fe412a req-8d21566f-63a5-49e6-9a2b-944055efbc7b service nova] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Refreshing instance network info cache due to event network-changed-99fff1a2-f410-40f3-9e32-fa6bb307a34a. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 920.510344] env[61243]: DEBUG oslo_concurrency.lockutils [req-ea11618d-616e-4111-822a-bfee05fe412a req-8d21566f-63a5-49e6-9a2b-944055efbc7b service nova] Acquiring lock "refresh_cache-09587d68-f7fd-4648-bd5c-e53b1b6464e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.573464] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.625883] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339115, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.706722] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a42ddc10-3475-4fec-a962-f6c0ec8172e3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.715241] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-326336b2-c59f-4cac-b9a4-ca05fb672e3b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.768434] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "894eb798-6c9e-47cb-8eb5-4610fe184bc5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.768895] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "894eb798-6c9e-47cb-8eb5-4610fe184bc5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.769267] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "894eb798-6c9e-47cb-8eb5-4610fe184bc5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.769624] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "894eb798-6c9e-47cb-8eb5-4610fe184bc5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.769922] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "894eb798-6c9e-47cb-8eb5-4610fe184bc5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.778142] env[61243]: INFO nova.compute.manager [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Terminating instance [ 920.781540] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bb7627-4147-41ed-beec-72188f5bcefe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.786460] env[61243]: DEBUG nova.compute.manager [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 920.786858] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 920.787474] env[61243]: DEBUG oslo_concurrency.lockutils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Releasing lock "refresh_cache-09587d68-f7fd-4648-bd5c-e53b1b6464e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.788134] env[61243]: DEBUG nova.compute.manager [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Instance network_info: |[{"id": "99fff1a2-f410-40f3-9e32-fa6bb307a34a", "address": "fa:16:3e:44:ae:bb", "network": {"id": "2737f3f1-a8da-4b5f-a33d-c768f14207e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1004846100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3248cca8a8aa4cad84b8bd5dcf7ed0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99fff1a2-f4", "ovs_interfaceid": "99fff1a2-f410-40f3-9e32-fa6bb307a34a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 920.789061] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-097870b5-ce00-456c-b0a6-ecb844758b66 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.794409] env[61243]: DEBUG oslo_concurrency.lockutils [req-ea11618d-616e-4111-822a-bfee05fe412a req-8d21566f-63a5-49e6-9a2b-944055efbc7b service nova] Acquired lock "refresh_cache-09587d68-f7fd-4648-bd5c-e53b1b6464e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.794759] env[61243]: DEBUG nova.network.neutron [req-ea11618d-616e-4111-822a-bfee05fe412a req-8d21566f-63a5-49e6-9a2b-944055efbc7b service nova] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Refreshing network info cache for port 99fff1a2-f410-40f3-9e32-fa6bb307a34a {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 920.796516] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:ae:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '26472e27-9835-4f87-ab7f-ca24dfee4e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '99fff1a2-f410-40f3-9e32-fa6bb307a34a', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 920.804636] env[61243]: DEBUG oslo.service.loopingcall [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.807445] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 920.808735] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-917efe2b-78b3-4a62-8ffd-80bb2d4f14fe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.825082] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-995fc36a-3199-4364-9556-c08840e19942 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.832240] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 920.835965] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e293033-7b2b-4808-83f7-dbd98fd04b9e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.846245] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 920.846245] env[61243]: value = "task-1339116" [ 920.846245] env[61243]: _type = "Task" [ 920.846245] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.846785] env[61243]: DEBUG nova.compute.provider_tree [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.856525] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e0c642-eed3-b8c2-7815-80d3bd3763af, 'name': SearchDatastore_Task, 'duration_secs': 0.018455} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.858329] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.858604] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 5c063b4a-3fbb-4222-ab6d-7d27b6b99084/5c063b4a-3fbb-4222-ab6d-7d27b6b99084.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 920.859305] env[61243]: DEBUG oslo_vmware.api [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 920.859305] env[61243]: value = "task-1339117" [ 920.859305] env[61243]: _type = "Task" [ 920.859305] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.859531] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07cce94c-bb5f-4337-9139-714ce2ab4fe8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.868159] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339116, 'name': CreateVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.874557] env[61243]: DEBUG oslo_vmware.api [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339117, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.876381] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 920.876381] env[61243]: value = "task-1339118" [ 920.876381] env[61243]: _type = "Task" [ 920.876381] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.886428] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339118, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.964235] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339109, 'name': CloneVM_Task, 'duration_secs': 1.631928} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.964838] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-093ad0fe-8eb7-42a3-9348-e903beb275aa tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 920.965151] env[61243]: INFO nova.virt.vmwareapi.vmops [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Created linked-clone VM from snapshot [ 920.965419] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-12c59234-37dd-4814-b818-059f6f0a4115 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.968483] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448308fd-f79b-4d4f-8ddf-fb936b4aa762 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.977466] env[61243]: DEBUG oslo_vmware.api [None req-093ad0fe-8eb7-42a3-9348-e903beb275aa tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 920.977466] env[61243]: value = "task-1339119" [ 920.977466] env[61243]: _type = "Task" [ 920.977466] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.980895] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Uploading image f26780ef-1234-46b6-a400-0b3f1542aed9 {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 920.994167] env[61243]: DEBUG oslo_vmware.api [None req-093ad0fe-8eb7-42a3-9348-e903beb275aa tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339119, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.016059] env[61243]: DEBUG oslo_vmware.rw_handles [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 921.016059] env[61243]: value = "vm-285742" [ 921.016059] env[61243]: _type = "VirtualMachine" [ 921.016059] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 921.016059] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c3285187-9eb8-4591-b3c4-64f0e18d82db {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.024647] env[61243]: DEBUG oslo_vmware.rw_handles [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lease: (returnval){ [ 921.024647] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529f61e3-1c9c-33bf-0f9f-6f2599123c95" [ 921.024647] env[61243]: _type = "HttpNfcLease" [ 921.024647] env[61243]: } obtained for exporting VM: (result){ [ 921.024647] env[61243]: value = "vm-285742" [ 921.024647] env[61243]: _type = "VirtualMachine" [ 921.024647] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 921.025874] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the lease: (returnval){ [ 921.025874] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529f61e3-1c9c-33bf-0f9f-6f2599123c95" [ 921.025874] env[61243]: _type = "HttpNfcLease" [ 921.025874] env[61243]: } to be ready. {{(pid=61243) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 921.032644] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 921.032644] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529f61e3-1c9c-33bf-0f9f-6f2599123c95" [ 921.032644] env[61243]: _type = "HttpNfcLease" [ 921.032644] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 921.079127] env[61243]: DEBUG oslo_vmware.rw_handles [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f80849-2c16-7302-5e76-bc91067b1eeb/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 921.080168] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad954fa-4d5d-47ab-aa45-b894fee6c642 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.088889] env[61243]: DEBUG oslo_vmware.rw_handles [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f80849-2c16-7302-5e76-bc91067b1eeb/disk-0.vmdk is in state: ready. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 921.089106] env[61243]: ERROR oslo_vmware.rw_handles [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f80849-2c16-7302-5e76-bc91067b1eeb/disk-0.vmdk due to incomplete transfer. [ 921.089595] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b02179bf-9148-4a37-b72d-a50f2bc8fa5a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.097271] env[61243]: DEBUG oslo_vmware.rw_handles [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f80849-2c16-7302-5e76-bc91067b1eeb/disk-0.vmdk. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 921.097554] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Uploaded image 0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e to the Glance image server {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 921.100074] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Destroying the VM {{(pid=61243) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 921.100413] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e1e1e5d3-cc1f-4341-a4ef-242f34941c8d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.114048] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 921.114048] env[61243]: value = "task-1339121" [ 921.114048] env[61243]: _type = "Task" [ 921.114048] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.127594] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339121, 'name': Destroy_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.131020] env[61243]: DEBUG oslo_vmware.api [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339115, 'name': PowerOnVM_Task, 'duration_secs': 0.784311} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.131382] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 921.131662] env[61243]: INFO nova.compute.manager [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Took 9.11 seconds to spawn the instance on the hypervisor. [ 921.131825] env[61243]: DEBUG nova.compute.manager [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 921.132676] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df21c1b7-10b9-4816-b616-e49e3fdb61cd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.316028] env[61243]: DEBUG nova.virt.hardware [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 921.316028] env[61243]: DEBUG nova.virt.hardware [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 921.316028] env[61243]: DEBUG nova.virt.hardware [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 921.316028] env[61243]: DEBUG nova.virt.hardware [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 921.316028] env[61243]: DEBUG nova.virt.hardware [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 921.316028] env[61243]: DEBUG nova.virt.hardware [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 921.316028] env[61243]: DEBUG nova.virt.hardware [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 921.316028] env[61243]: DEBUG nova.virt.hardware [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 921.316028] env[61243]: DEBUG nova.virt.hardware [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 921.316028] env[61243]: DEBUG nova.virt.hardware [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 921.316028] env[61243]: DEBUG nova.virt.hardware [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 921.317576] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c631a9-b744-41a4-8b6e-933eb35488e9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.328009] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbfef003-9f3b-4e1e-b36f-480625d66371 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.344679] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:51:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a071ecf4-e713-4f97-9271-8c17952f6dee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 921.350195] env[61243]: DEBUG oslo.service.loopingcall [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.351337] env[61243]: DEBUG nova.scheduler.client.report [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 921.354747] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 921.358217] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ceb7203d-104b-4e18-aad7-c550c9458af7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.391523] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339116, 'name': CreateVM_Task, 'duration_secs': 0.406308} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.393495] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 921.393973] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 921.393973] env[61243]: value = "task-1339122" [ 921.393973] env[61243]: _type = "Task" [ 921.393973] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.395262] env[61243]: DEBUG oslo_concurrency.lockutils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2b9c3b0a-5f89-462e-99b2-0c88500ad15b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.395262] env[61243]: DEBUG oslo_concurrency.lockutils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2b9c3b0a-5f89-462e-99b2-0c88500ad15b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.395566] env[61243]: DEBUG oslo_concurrency.lockutils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2b9c3b0a-5f89-462e-99b2-0c88500ad15b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 921.399240] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf07e4aa-6882-4396-93a6-742f144084d2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.404981] env[61243]: DEBUG oslo_vmware.api [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339117, 'name': PowerOffVM_Task, 'duration_secs': 0.286026} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.408567] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.408762] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 921.409473] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339118, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.410405] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a233155-48e7-4d61-9637-9ff25709f436 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.414473] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 921.414473] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52b0b29c-05d4-1a9f-42b0-3f3e6194be7b" [ 921.414473] env[61243]: _type = "Task" [ 921.414473] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.419674] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339122, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.431851] env[61243]: DEBUG oslo_concurrency.lockutils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2b9c3b0a-5f89-462e-99b2-0c88500ad15b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.432936] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Processing image 2b9c3b0a-5f89-462e-99b2-0c88500ad15b {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 921.433310] env[61243]: DEBUG oslo_concurrency.lockutils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2b9c3b0a-5f89-462e-99b2-0c88500ad15b/2b9c3b0a-5f89-462e-99b2-0c88500ad15b.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.434183] env[61243]: DEBUG oslo_concurrency.lockutils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2b9c3b0a-5f89-462e-99b2-0c88500ad15b/2b9c3b0a-5f89-462e-99b2-0c88500ad15b.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.434183] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 921.434183] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-37bdae84-ab8b-4bdc-8da3-d5f127ad9156 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.457470] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 921.457804] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 921.458916] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e323de12-e539-4e8a-9959-845f9a33396e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.466422] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 921.466422] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529b2181-2b4a-0cf9-fba2-b212a3ea4d59" [ 921.466422] env[61243]: _type = "Task" [ 921.466422] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.478422] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529b2181-2b4a-0cf9-fba2-b212a3ea4d59, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.490689] env[61243]: DEBUG oslo_vmware.api [None req-093ad0fe-8eb7-42a3-9348-e903beb275aa tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339119, 'name': PowerOffVM_Task, 'duration_secs': 0.322122} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.491017] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-093ad0fe-8eb7-42a3-9348-e903beb275aa tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.491213] env[61243]: DEBUG nova.compute.manager [None req-093ad0fe-8eb7-42a3-9348-e903beb275aa tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 921.492068] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d594d3-ce2d-4bdc-9da7-f289ea886bab {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.503615] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 921.503856] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 921.504050] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Deleting the datastore file [datastore2] 894eb798-6c9e-47cb-8eb5-4610fe184bc5 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.504319] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-491ab657-f04f-415a-8a8c-09d74b886feb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.510374] env[61243]: DEBUG oslo_vmware.api [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for the task: (returnval){ [ 921.510374] env[61243]: value = "task-1339124" [ 921.510374] env[61243]: _type = "Task" [ 921.510374] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.521240] env[61243]: DEBUG oslo_vmware.api [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339124, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.532935] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 921.532935] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529f61e3-1c9c-33bf-0f9f-6f2599123c95" [ 921.532935] env[61243]: _type = "HttpNfcLease" [ 921.532935] env[61243]: } is ready. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 921.533299] env[61243]: DEBUG oslo_vmware.rw_handles [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 921.533299] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529f61e3-1c9c-33bf-0f9f-6f2599123c95" [ 921.533299] env[61243]: _type = "HttpNfcLease" [ 921.533299] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 921.534089] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a3dee98-55bc-447c-8d4a-7fcaff053a3e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.544023] env[61243]: DEBUG oslo_vmware.rw_handles [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52edd12d-ba48-558c-a422-062c68f15f84/disk-0.vmdk from lease info. {{(pid=61243) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 921.544023] env[61243]: DEBUG oslo_vmware.rw_handles [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52edd12d-ba48-558c-a422-062c68f15f84/disk-0.vmdk for reading. {{(pid=61243) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 921.623438] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339121, 'name': Destroy_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.637304] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b87670d4-2dc4-4caf-8418-33a06526e459 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.652905] env[61243]: DEBUG nova.network.neutron [req-ea11618d-616e-4111-822a-bfee05fe412a req-8d21566f-63a5-49e6-9a2b-944055efbc7b service nova] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Updated VIF entry in instance network info cache for port 99fff1a2-f410-40f3-9e32-fa6bb307a34a. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 921.653078] env[61243]: DEBUG nova.network.neutron [req-ea11618d-616e-4111-822a-bfee05fe412a req-8d21566f-63a5-49e6-9a2b-944055efbc7b service nova] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Updating instance_info_cache with network_info: [{"id": "99fff1a2-f410-40f3-9e32-fa6bb307a34a", "address": "fa:16:3e:44:ae:bb", "network": {"id": "2737f3f1-a8da-4b5f-a33d-c768f14207e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1004846100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3248cca8a8aa4cad84b8bd5dcf7ed0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99fff1a2-f4", "ovs_interfaceid": "99fff1a2-f410-40f3-9e32-fa6bb307a34a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.660167] env[61243]: INFO nova.compute.manager [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Took 27.81 seconds to build instance. [ 921.855980] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.924s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.859062] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.058s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.859062] env[61243]: DEBUG nova.objects.instance [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lazy-loading 'resources' on Instance uuid d95a4d61-2bff-4d89-9ab8-f71e558f3409 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.880886] env[61243]: INFO nova.scheduler.client.report [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Deleted allocations for instance f8816b86-10c8-4d64-9659-fbc301893d46 [ 921.893821] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339118, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.609378} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.894122] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 5c063b4a-3fbb-4222-ab6d-7d27b6b99084/5c063b4a-3fbb-4222-ab6d-7d27b6b99084.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 921.894341] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 921.894637] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-869ee3fe-cd26-47c7-b8d5-833ab9ee1b2d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.906372] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339122, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.908110] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 921.908110] env[61243]: value = "task-1339125" [ 921.908110] env[61243]: _type = "Task" [ 921.908110] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.922176] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339125, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.978282] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Preparing fetch location {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 921.978611] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Fetch image to [datastore1] OSTACK_IMG_cbdd30f6-3502-48db-be2c-d4b11e96396c/OSTACK_IMG_cbdd30f6-3502-48db-be2c-d4b11e96396c.vmdk {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 921.978800] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Downloading stream optimized image 2b9c3b0a-5f89-462e-99b2-0c88500ad15b to [datastore1] OSTACK_IMG_cbdd30f6-3502-48db-be2c-d4b11e96396c/OSTACK_IMG_cbdd30f6-3502-48db-be2c-d4b11e96396c.vmdk on the data store datastore1 as vApp {{(pid=61243) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 921.979012] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Downloading image file data 2b9c3b0a-5f89-462e-99b2-0c88500ad15b to the ESX as VM named 'OSTACK_IMG_cbdd30f6-3502-48db-be2c-d4b11e96396c' {{(pid=61243) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 922.031141] env[61243]: DEBUG oslo_concurrency.lockutils [None req-093ad0fe-8eb7-42a3-9348-e903beb275aa tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "ebe8cc8e-cc35-4ec4-bed7-d55df72a2991" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.587s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.039811] env[61243]: DEBUG oslo_vmware.api [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339124, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.079738] env[61243]: DEBUG oslo_vmware.rw_handles [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 922.079738] env[61243]: value = "resgroup-9" [ 922.079738] env[61243]: _type = "ResourcePool" [ 922.079738] env[61243]: }. {{(pid=61243) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 922.080583] env[61243]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-6c03261c-6eba-4192-98f6-de57c5f0ca05 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.101245] env[61243]: DEBUG oslo_vmware.rw_handles [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lease: (returnval){ [ 922.101245] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521236c5-c448-3be1-e900-2d42484fdf27" [ 922.101245] env[61243]: _type = "HttpNfcLease" [ 922.101245] env[61243]: } obtained for vApp import into resource pool (val){ [ 922.101245] env[61243]: value = "resgroup-9" [ 922.101245] env[61243]: _type = "ResourcePool" [ 922.101245] env[61243]: }. {{(pid=61243) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 922.101741] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the lease: (returnval){ [ 922.101741] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521236c5-c448-3be1-e900-2d42484fdf27" [ 922.101741] env[61243]: _type = "HttpNfcLease" [ 922.101741] env[61243]: } to be ready. {{(pid=61243) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 922.110268] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 922.110268] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521236c5-c448-3be1-e900-2d42484fdf27" [ 922.110268] env[61243]: _type = "HttpNfcLease" [ 922.110268] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 922.125365] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339121, 'name': Destroy_Task, 'duration_secs': 0.682736} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.125365] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Destroyed the VM [ 922.125365] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Deleting Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 922.125965] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-98cb2764-7089-4c4a-8402-0b2032ae8363 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.132784] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 922.132784] env[61243]: value = "task-1339127" [ 922.132784] env[61243]: _type = "Task" [ 922.132784] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.141685] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339127, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.160220] env[61243]: DEBUG oslo_concurrency.lockutils [req-ea11618d-616e-4111-822a-bfee05fe412a req-8d21566f-63a5-49e6-9a2b-944055efbc7b service nova] Releasing lock "refresh_cache-09587d68-f7fd-4648-bd5c-e53b1b6464e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.160220] env[61243]: DEBUG nova.compute.manager [req-ea11618d-616e-4111-822a-bfee05fe412a req-8d21566f-63a5-49e6-9a2b-944055efbc7b service nova] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Received event network-vif-deleted-5ae039d5-ce2c-4973-bea6-c6e02dfd1ca7 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.160220] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2b96a93a-a68b-4fc4-aad5-82c806a714cf tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d46fb6dc-cdd9-4688-b5e7-9efde0f770c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.322s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.395816] env[61243]: DEBUG oslo_concurrency.lockutils [None req-93e726a7-bfec-4cc2-a006-3fdea883d028 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "f8816b86-10c8-4d64-9659-fbc301893d46" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.960s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.410386] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339122, 'name': CreateVM_Task, 'duration_secs': 0.691559} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.413922] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 922.415417] env[61243]: DEBUG oslo_concurrency.lockutils [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.415603] env[61243]: DEBUG oslo_concurrency.lockutils [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.416153] env[61243]: DEBUG oslo_concurrency.lockutils [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 922.417617] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e04990c8-6833-4027-9e13-835a56edf0e1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.423280] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339125, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084747} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.426801] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 922.427830] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da97711a-a1f3-4794-abef-26afab585878 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.432842] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 922.432842] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524732c0-fe48-770c-47c0-107ee60b0d07" [ 922.432842] env[61243]: _type = "Task" [ 922.432842] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.462147] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] 5c063b4a-3fbb-4222-ab6d-7d27b6b99084/5c063b4a-3fbb-4222-ab6d-7d27b6b99084.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.463588] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ebf4a35-a74a-4a83-8797-22b89f3a5b0a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.488178] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524732c0-fe48-770c-47c0-107ee60b0d07, 'name': SearchDatastore_Task, 'duration_secs': 0.016064} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.489443] env[61243]: DEBUG oslo_concurrency.lockutils [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.489791] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 922.490056] env[61243]: DEBUG oslo_concurrency.lockutils [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.490224] env[61243]: DEBUG oslo_concurrency.lockutils [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.490421] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 922.490706] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-73ad7f5e-3802-4c5f-b1d1-288f77447074 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.494915] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 922.494915] env[61243]: value = "task-1339128" [ 922.494915] env[61243]: _type = "Task" [ 922.494915] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.506900] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339128, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.507201] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 922.507386] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 922.508299] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e03af660-66a4-4cd2-a8a0-05a1dd1f7842 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.514241] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 922.514241] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525b6706-2131-4842-ff83-82fe2a65cefa" [ 922.514241] env[61243]: _type = "Task" [ 922.514241] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.525601] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525b6706-2131-4842-ff83-82fe2a65cefa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.539041] env[61243]: DEBUG oslo_vmware.api [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Task: {'id': task-1339124, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.542088} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.539398] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.539648] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 922.539906] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 922.540097] env[61243]: INFO nova.compute.manager [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Took 1.75 seconds to destroy the instance on the hypervisor. [ 922.540332] env[61243]: DEBUG oslo.service.loopingcall [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.543328] env[61243]: DEBUG nova.compute.manager [-] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 922.543439] env[61243]: DEBUG nova.network.neutron [-] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 922.612163] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 922.612163] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521236c5-c448-3be1-e900-2d42484fdf27" [ 922.612163] env[61243]: _type = "HttpNfcLease" [ 922.612163] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 922.643612] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339127, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.701814] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a774a837-195a-4aa4-9625-94b9dc441a47 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.710303] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86076b42-6305-4389-901e-3ba5c6e87753 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.747609] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f7619ee-5cdd-4001-ac68-0c4ca802b0fd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.758045] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d8f202a-7dd8-47f7-a06b-d27c99df3673 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.777966] env[61243]: DEBUG nova.compute.provider_tree [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.981401] env[61243]: DEBUG nova.compute.manager [req-5fb2f35e-cd23-4507-81f5-7702561af564 req-7c75f50c-32cc-4309-99b0-0327ae06996e service nova] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Received event network-vif-deleted-bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.981970] env[61243]: INFO nova.compute.manager [req-5fb2f35e-cd23-4507-81f5-7702561af564 req-7c75f50c-32cc-4309-99b0-0327ae06996e service nova] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Neutron deleted interface bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0; detaching it from the instance and deleting it from the info cache [ 922.981970] env[61243]: DEBUG nova.network.neutron [req-5fb2f35e-cd23-4507-81f5-7702561af564 req-7c75f50c-32cc-4309-99b0-0327ae06996e service nova] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.007494] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339128, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.028645] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525b6706-2131-4842-ff83-82fe2a65cefa, 'name': SearchDatastore_Task, 'duration_secs': 0.023171} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.030291] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfd8ea09-b110-4117-80da-2eb3e48669b6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.039165] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 923.039165] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527cba72-9a06-6321-a7a0-be85da19d0a4" [ 923.039165] env[61243]: _type = "Task" [ 923.039165] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.058486] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527cba72-9a06-6321-a7a0-be85da19d0a4, 'name': SearchDatastore_Task} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.058486] env[61243]: DEBUG oslo_concurrency.lockutils [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.058486] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] e9a1c8af-7129-4366-bda1-1c3cf40eeab1/e9a1c8af-7129-4366-bda1-1c3cf40eeab1.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 923.058755] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-14423bf1-f87e-4424-b765-27ba45295982 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.069141] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 923.069141] env[61243]: value = "task-1339129" [ 923.069141] env[61243]: _type = "Task" [ 923.069141] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.087202] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339129, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.114074] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 923.114074] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521236c5-c448-3be1-e900-2d42484fdf27" [ 923.114074] env[61243]: _type = "HttpNfcLease" [ 923.114074] env[61243]: } is ready. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 923.114757] env[61243]: DEBUG oslo_vmware.rw_handles [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 923.114757] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521236c5-c448-3be1-e900-2d42484fdf27" [ 923.114757] env[61243]: _type = "HttpNfcLease" [ 923.114757] env[61243]: }. {{(pid=61243) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 923.115098] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1fb3d80-9609-4c91-9f02-e3a9a312e79a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.127423] env[61243]: DEBUG oslo_vmware.rw_handles [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52527c27-5b38-9269-aa20-70c92c02d30e/disk-0.vmdk from lease info. {{(pid=61243) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 923.127423] env[61243]: DEBUG oslo_vmware.rw_handles [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52527c27-5b38-9269-aa20-70c92c02d30e/disk-0.vmdk. {{(pid=61243) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 923.203246] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1ae0323a-23b4-43a5-9647-331951f53fce {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.204027] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339127, 'name': RemoveSnapshot_Task, 'duration_secs': 0.838972} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.204410] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Deleted Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 923.204841] env[61243]: DEBUG nova.compute.manager [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.206247] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e5f06f-36cc-4ca5-895c-91a0173d1f61 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.282136] env[61243]: DEBUG nova.scheduler.client.report [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 923.339156] env[61243]: DEBUG nova.network.neutron [-] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.403502] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "ebe8cc8e-cc35-4ec4-bed7-d55df72a2991" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.404806] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "ebe8cc8e-cc35-4ec4-bed7-d55df72a2991" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.404806] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "ebe8cc8e-cc35-4ec4-bed7-d55df72a2991-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.407540] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "ebe8cc8e-cc35-4ec4-bed7-d55df72a2991-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.407540] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "ebe8cc8e-cc35-4ec4-bed7-d55df72a2991-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.412328] env[61243]: INFO nova.compute.manager [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Terminating instance [ 923.415754] env[61243]: DEBUG nova.compute.manager [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 923.416167] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 923.417774] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ed622b-4522-48e1-8e6d-ac32bc40b26e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.432570] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 923.435247] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a99276cd-727d-48bd-ade0-178d542873a0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.484980] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b0aad30d-c5b7-4129-bda1-9b406721998f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.506038] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc3cd03-1cdc-49ab-880b-aa9763b5baa6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.532472] env[61243]: DEBUG nova.compute.manager [req-124a84b2-9ee5-4ae5-b0ed-d2389dd6a844 req-7076f5e9-c412-422a-a46a-1f180e126110 service nova] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Received event network-changed-bd92d46a-e97e-4169-a89e-59191cce73f3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.532772] env[61243]: DEBUG nova.compute.manager [req-124a84b2-9ee5-4ae5-b0ed-d2389dd6a844 req-7076f5e9-c412-422a-a46a-1f180e126110 service nova] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Refreshing instance network info cache due to event network-changed-bd92d46a-e97e-4169-a89e-59191cce73f3. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 923.532925] env[61243]: DEBUG oslo_concurrency.lockutils [req-124a84b2-9ee5-4ae5-b0ed-d2389dd6a844 req-7076f5e9-c412-422a-a46a-1f180e126110 service nova] Acquiring lock "refresh_cache-d46fb6dc-cdd9-4688-b5e7-9efde0f770c9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.533179] env[61243]: DEBUG oslo_concurrency.lockutils [req-124a84b2-9ee5-4ae5-b0ed-d2389dd6a844 req-7076f5e9-c412-422a-a46a-1f180e126110 service nova] Acquired lock "refresh_cache-d46fb6dc-cdd9-4688-b5e7-9efde0f770c9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.533374] env[61243]: DEBUG nova.network.neutron [req-124a84b2-9ee5-4ae5-b0ed-d2389dd6a844 req-7076f5e9-c412-422a-a46a-1f180e126110 service nova] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Refreshing network info cache for port bd92d46a-e97e-4169-a89e-59191cce73f3 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 923.537537] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 923.537815] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 923.538046] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Deleting the datastore file [datastore2] ebe8cc8e-cc35-4ec4-bed7-d55df72a2991 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 923.538839] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a69eb1ba-ddc2-4bb2-9051-9ebe40074eac {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.545541] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339128, 'name': ReconfigVM_Task, 'duration_secs': 0.593663} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.550727] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Reconfigured VM instance instance-00000053 to attach disk [datastore2] 5c063b4a-3fbb-4222-ab6d-7d27b6b99084/5c063b4a-3fbb-4222-ab6d-7d27b6b99084.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.563661] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3a684c69-153f-4d96-87f8-2a494cd9183b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.566854] env[61243]: DEBUG nova.compute.manager [req-5fb2f35e-cd23-4507-81f5-7702561af564 req-7c75f50c-32cc-4309-99b0-0327ae06996e service nova] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Detach interface failed, port_id=bdfbe0d5-9b56-4434-a74e-0f9f8dfa13a0, reason: Instance 894eb798-6c9e-47cb-8eb5-4610fe184bc5 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 923.573206] env[61243]: DEBUG oslo_vmware.api [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 923.573206] env[61243]: value = "task-1339131" [ 923.573206] env[61243]: _type = "Task" [ 923.573206] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.585264] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 923.585264] env[61243]: value = "task-1339132" [ 923.585264] env[61243]: _type = "Task" [ 923.585264] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.601273] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339129, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.601576] env[61243]: DEBUG oslo_vmware.api [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339131, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.612783] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339132, 'name': Rename_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.724170] env[61243]: INFO nova.compute.manager [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Shelve offloading [ 923.727430] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 923.727827] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b0fe47d2-0b1f-4db4-8765-b8307262f1c0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.742274] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 923.742274] env[61243]: value = "task-1339133" [ 923.742274] env[61243]: _type = "Task" [ 923.742274] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.755379] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] VM already powered off {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 923.755698] env[61243]: DEBUG nova.compute.manager [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.756761] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1666144c-0d37-4928-86e7-270083d6c389 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.767845] env[61243]: DEBUG oslo_concurrency.lockutils [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.768127] env[61243]: DEBUG oslo_concurrency.lockutils [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquired lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.768276] env[61243]: DEBUG nova.network.neutron [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 923.788040] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.929s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.794934] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.416s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.797177] env[61243]: INFO nova.compute.claims [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 923.818843] env[61243]: INFO nova.scheduler.client.report [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Deleted allocations for instance d95a4d61-2bff-4d89-9ab8-f71e558f3409 [ 923.842699] env[61243]: INFO nova.compute.manager [-] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Took 1.30 seconds to deallocate network for instance. [ 924.037024] env[61243]: DEBUG oslo_vmware.rw_handles [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Completed reading data from the image iterator. {{(pid=61243) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 924.037024] env[61243]: DEBUG oslo_vmware.rw_handles [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52527c27-5b38-9269-aa20-70c92c02d30e/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 924.037024] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87fbc24c-3475-43c6-bbfa-415ce00797c2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.054196] env[61243]: DEBUG oslo_vmware.rw_handles [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52527c27-5b38-9269-aa20-70c92c02d30e/disk-0.vmdk is in state: ready. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 924.054958] env[61243]: DEBUG oslo_vmware.rw_handles [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52527c27-5b38-9269-aa20-70c92c02d30e/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 924.054958] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-cfb29583-5a5d-4800-9105-45e021e425c5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.090325] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339129, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.804617} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.098125] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] e9a1c8af-7129-4366-bda1-1c3cf40eeab1/e9a1c8af-7129-4366-bda1-1c3cf40eeab1.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 924.098400] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 924.098733] env[61243]: DEBUG oslo_vmware.api [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339131, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.335198} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.098956] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c7b80b08-5006-4d75-9459-406145bc04fb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.101039] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 924.101470] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 924.101721] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 924.101856] env[61243]: INFO nova.compute.manager [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Took 0.69 seconds to destroy the instance on the hypervisor. [ 924.102166] env[61243]: DEBUG oslo.service.loopingcall [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.102806] env[61243]: DEBUG nova.compute.manager [-] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 924.102921] env[61243]: DEBUG nova.network.neutron [-] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 924.109173] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339132, 'name': Rename_Task, 'duration_secs': 0.268139} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.110700] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 924.111334] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 924.111334] env[61243]: value = "task-1339134" [ 924.111334] env[61243]: _type = "Task" [ 924.111334] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.111334] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bbd5d5f8-dcb0-4a50-9ac1-9393e61cc735 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.124218] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339134, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.125877] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 924.125877] env[61243]: value = "task-1339135" [ 924.125877] env[61243]: _type = "Task" [ 924.125877] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.136763] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339135, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.231285] env[61243]: DEBUG oslo_vmware.rw_handles [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52527c27-5b38-9269-aa20-70c92c02d30e/disk-0.vmdk. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 924.231515] env[61243]: INFO nova.virt.vmwareapi.images [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Downloaded image file data 2b9c3b0a-5f89-462e-99b2-0c88500ad15b [ 924.232467] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d1a8fc-6313-446c-9cd9-75c6b0160c53 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.256779] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d81cad4c-71be-4a01-9d08-54fa91aa62aa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.293180] env[61243]: INFO nova.virt.vmwareapi.images [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] The imported VM was unregistered [ 924.295496] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Caching image {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 924.295799] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Creating directory with path [datastore1] devstack-image-cache_base/2b9c3b0a-5f89-462e-99b2-0c88500ad15b {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 924.296121] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b11d3657-b8ba-45be-9768-b48e4f36f057 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.321024] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Created directory with path [datastore1] devstack-image-cache_base/2b9c3b0a-5f89-462e-99b2-0c88500ad15b {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 924.321024] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_cbdd30f6-3502-48db-be2c-d4b11e96396c/OSTACK_IMG_cbdd30f6-3502-48db-be2c-d4b11e96396c.vmdk to [datastore1] devstack-image-cache_base/2b9c3b0a-5f89-462e-99b2-0c88500ad15b/2b9c3b0a-5f89-462e-99b2-0c88500ad15b.vmdk. {{(pid=61243) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 924.321024] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-46f5580b-997f-4a1a-8ce8-a3bafeca616c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.328465] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 924.328465] env[61243]: value = "task-1339137" [ 924.328465] env[61243]: _type = "Task" [ 924.328465] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.337878] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6f81b2b9-0f34-4520-9b0f-f144225a5715 tempest-MultipleCreateTestJSON-762350172 tempest-MultipleCreateTestJSON-762350172-project-member] Lock "d95a4d61-2bff-4d89-9ab8-f71e558f3409" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.629s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.347858] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339137, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.360183] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.559777] env[61243]: DEBUG nova.network.neutron [req-124a84b2-9ee5-4ae5-b0ed-d2389dd6a844 req-7076f5e9-c412-422a-a46a-1f180e126110 service nova] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Updated VIF entry in instance network info cache for port bd92d46a-e97e-4169-a89e-59191cce73f3. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 924.560255] env[61243]: DEBUG nova.network.neutron [req-124a84b2-9ee5-4ae5-b0ed-d2389dd6a844 req-7076f5e9-c412-422a-a46a-1f180e126110 service nova] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Updating instance_info_cache with network_info: [{"id": "bd92d46a-e97e-4169-a89e-59191cce73f3", "address": "fa:16:3e:97:d0:81", "network": {"id": "523d4cf1-467c-4ea6-80a4-ebbd01016efa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-919325376-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7651b565d1647eb948fe07faba32a41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd92d46a-e9", "ovs_interfaceid": "bd92d46a-e97e-4169-a89e-59191cce73f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.624532] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339134, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094081} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.625036] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 924.625782] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dce6f06f-700b-4dd4-8462-0252a4130290 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.640286] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339135, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.666609] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] e9a1c8af-7129-4366-bda1-1c3cf40eeab1/e9a1c8af-7129-4366-bda1-1c3cf40eeab1.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 924.670569] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f8287bc-4e07-4238-989f-bfdd11633de6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.695123] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 924.695123] env[61243]: value = "task-1339138" [ 924.695123] env[61243]: _type = "Task" [ 924.695123] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.708858] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339138, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.736625] env[61243]: DEBUG nova.network.neutron [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Updating instance_info_cache with network_info: [{"id": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "address": "fa:16:3e:b0:f1:66", "network": {"id": "49be247d-97e7-4085-8272-4ccf41767d10", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1335943921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "399030c8863346ea97f6da669ff3868b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2248ec9-02", "ovs_interfaceid": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.841161] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339137, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.015730] env[61243]: DEBUG nova.compute.manager [req-fb4c3d16-86f3-47a1-9ee3-7360b590ab7a req-32d7b220-41f2-49df-96c1-9f65c8f4ce2e service nova] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Received event network-vif-deleted-f3206fdf-946a-4d55-b5a7-3b909da22db2 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.016378] env[61243]: INFO nova.compute.manager [req-fb4c3d16-86f3-47a1-9ee3-7360b590ab7a req-32d7b220-41f2-49df-96c1-9f65c8f4ce2e service nova] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Neutron deleted interface f3206fdf-946a-4d55-b5a7-3b909da22db2; detaching it from the instance and deleting it from the info cache [ 925.016769] env[61243]: DEBUG nova.network.neutron [req-fb4c3d16-86f3-47a1-9ee3-7360b590ab7a req-32d7b220-41f2-49df-96c1-9f65c8f4ce2e service nova] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.062910] env[61243]: DEBUG oslo_concurrency.lockutils [req-124a84b2-9ee5-4ae5-b0ed-d2389dd6a844 req-7076f5e9-c412-422a-a46a-1f180e126110 service nova] Releasing lock "refresh_cache-d46fb6dc-cdd9-4688-b5e7-9efde0f770c9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.077884] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f3d489-ebe4-438e-87f5-fc4bef1fa433 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.090197] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b05e54e-f0c3-4184-9f29-82cc4a8fd6b6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.138993] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ad88b0-8c66-4940-b4b6-3dd7c510513f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.150100] env[61243]: DEBUG oslo_vmware.api [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339135, 'name': PowerOnVM_Task, 'duration_secs': 0.758415} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.152249] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 925.152430] env[61243]: INFO nova.compute.manager [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Took 9.58 seconds to spawn the instance on the hypervisor. [ 925.152708] env[61243]: DEBUG nova.compute.manager [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.153729] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1346cbb6-bbb9-463a-bfb6-3d64a974c303 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.158628] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bfea040-43e7-4f9e-98a0-6c686c15bf45 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.175975] env[61243]: DEBUG nova.compute.provider_tree [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 925.192289] env[61243]: DEBUG nova.network.neutron [-] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.209998] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339138, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.240469] env[61243]: DEBUG oslo_concurrency.lockutils [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Releasing lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.342580] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339137, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.519777] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cccfb343-dd35-4658-8f57-3324c4ab6995 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.534511] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4197f92-7234-416c-920c-6814a04106c2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.574272] env[61243]: DEBUG nova.compute.manager [req-fb4c3d16-86f3-47a1-9ee3-7360b590ab7a req-32d7b220-41f2-49df-96c1-9f65c8f4ce2e service nova] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Detach interface failed, port_id=f3206fdf-946a-4d55-b5a7-3b909da22db2, reason: Instance ebe8cc8e-cc35-4ec4-bed7-d55df72a2991 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 925.595610] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 925.596978] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08374cea-dd91-48d1-92d9-09f0db887176 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.612443] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 925.612898] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e6c5da7-0202-496a-8649-153248bf183d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.699451] env[61243]: INFO nova.compute.manager [-] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Took 1.60 seconds to deallocate network for instance. [ 925.711990] env[61243]: INFO nova.compute.manager [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Took 30.84 seconds to build instance. [ 925.722033] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339138, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.724909] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 925.725182] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 925.725370] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Deleting the datastore file [datastore2] 86179fb0-99df-4b10-a815-c19168e9521e {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.725667] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc394e0c-0485-44f7-9f96-e01b6938de30 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.729950] env[61243]: ERROR nova.scheduler.client.report [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [req-7efeff6d-a6d6-48e6-bfd4-bf7b42488f72] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7efeff6d-a6d6-48e6-bfd4-bf7b42488f72"}]} [ 925.735848] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 925.735848] env[61243]: value = "task-1339140" [ 925.735848] env[61243]: _type = "Task" [ 925.735848] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.745942] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.754877] env[61243]: DEBUG nova.scheduler.client.report [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 925.772458] env[61243]: DEBUG nova.scheduler.client.report [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 925.772765] env[61243]: DEBUG nova.compute.provider_tree [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 925.785903] env[61243]: DEBUG nova.scheduler.client.report [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 925.809136] env[61243]: DEBUG nova.scheduler.client.report [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 925.846153] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339137, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.167377] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad9e7dd5-53c9-4fba-8337-2a227d9d1bcc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.178807] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0db25f-aaeb-410f-aa39-54286533fe32 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.217287] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3288484-4a07-4ac1-befe-6b76a6e5769d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.221021] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.221452] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4e984148-e999-4078-bbfd-7819a8524ec6 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "5c063b4a-3fbb-4222-ab6d-7d27b6b99084" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.360s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.232614] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02785290-c092-4b8c-9eff-2f0793f4c2d8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.237127] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339138, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.247780] env[61243]: DEBUG oslo_vmware.api [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339140, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.344402} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.258679] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 926.258679] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 926.258679] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 926.260604] env[61243]: DEBUG nova.compute.provider_tree [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 926.287634] env[61243]: INFO nova.scheduler.client.report [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Deleted allocations for instance 86179fb0-99df-4b10-a815-c19168e9521e [ 926.347637] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339137, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.725727] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339138, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.780491] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "5c063b4a-3fbb-4222-ab6d-7d27b6b99084" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.780768] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "5c063b4a-3fbb-4222-ab6d-7d27b6b99084" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.780973] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "5c063b4a-3fbb-4222-ab6d-7d27b6b99084-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.781200] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "5c063b4a-3fbb-4222-ab6d-7d27b6b99084-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.781356] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "5c063b4a-3fbb-4222-ab6d-7d27b6b99084-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.784139] env[61243]: INFO nova.compute.manager [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Terminating instance [ 926.786090] env[61243]: DEBUG nova.compute.manager [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 926.786261] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 926.787261] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9516b25b-dd66-4cd2-8a1e-70f691748077 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.793574] env[61243]: DEBUG oslo_concurrency.lockutils [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.796469] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 926.796892] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6b4d3440-3956-4845-80ba-3e0ad58a8b1d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.802673] env[61243]: DEBUG nova.scheduler.client.report [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 111 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 926.802975] env[61243]: DEBUG nova.compute.provider_tree [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 111 to 112 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 926.803246] env[61243]: DEBUG nova.compute.provider_tree [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 926.809064] env[61243]: DEBUG oslo_vmware.api [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 926.809064] env[61243]: value = "task-1339141" [ 926.809064] env[61243]: _type = "Task" [ 926.809064] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.818454] env[61243]: DEBUG oslo_vmware.api [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339141, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.847378] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339137, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.114444] env[61243]: DEBUG nova.compute.manager [req-29fe67de-ec5d-4df9-98eb-ae8c8c27486e req-3e6d63cc-8a57-4972-bafe-b2ebce216a29 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Received event network-vif-unplugged-f2248ec9-0230-4aa8-a016-a30ddf4f45f4 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.114444] env[61243]: DEBUG oslo_concurrency.lockutils [req-29fe67de-ec5d-4df9-98eb-ae8c8c27486e req-3e6d63cc-8a57-4972-bafe-b2ebce216a29 service nova] Acquiring lock "86179fb0-99df-4b10-a815-c19168e9521e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.114583] env[61243]: DEBUG oslo_concurrency.lockutils [req-29fe67de-ec5d-4df9-98eb-ae8c8c27486e req-3e6d63cc-8a57-4972-bafe-b2ebce216a29 service nova] Lock "86179fb0-99df-4b10-a815-c19168e9521e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.114768] env[61243]: DEBUG oslo_concurrency.lockutils [req-29fe67de-ec5d-4df9-98eb-ae8c8c27486e req-3e6d63cc-8a57-4972-bafe-b2ebce216a29 service nova] Lock "86179fb0-99df-4b10-a815-c19168e9521e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.114957] env[61243]: DEBUG nova.compute.manager [req-29fe67de-ec5d-4df9-98eb-ae8c8c27486e req-3e6d63cc-8a57-4972-bafe-b2ebce216a29 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] No waiting events found dispatching network-vif-unplugged-f2248ec9-0230-4aa8-a016-a30ddf4f45f4 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 927.115175] env[61243]: WARNING nova.compute.manager [req-29fe67de-ec5d-4df9-98eb-ae8c8c27486e req-3e6d63cc-8a57-4972-bafe-b2ebce216a29 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Received unexpected event network-vif-unplugged-f2248ec9-0230-4aa8-a016-a30ddf4f45f4 for instance with vm_state shelved_offloaded and task_state None. [ 927.115357] env[61243]: DEBUG nova.compute.manager [req-29fe67de-ec5d-4df9-98eb-ae8c8c27486e req-3e6d63cc-8a57-4972-bafe-b2ebce216a29 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Received event network-changed-f2248ec9-0230-4aa8-a016-a30ddf4f45f4 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.115590] env[61243]: DEBUG nova.compute.manager [req-29fe67de-ec5d-4df9-98eb-ae8c8c27486e req-3e6d63cc-8a57-4972-bafe-b2ebce216a29 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Refreshing instance network info cache due to event network-changed-f2248ec9-0230-4aa8-a016-a30ddf4f45f4. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 927.115823] env[61243]: DEBUG oslo_concurrency.lockutils [req-29fe67de-ec5d-4df9-98eb-ae8c8c27486e req-3e6d63cc-8a57-4972-bafe-b2ebce216a29 service nova] Acquiring lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.115978] env[61243]: DEBUG oslo_concurrency.lockutils [req-29fe67de-ec5d-4df9-98eb-ae8c8c27486e req-3e6d63cc-8a57-4972-bafe-b2ebce216a29 service nova] Acquired lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.116171] env[61243]: DEBUG nova.network.neutron [req-29fe67de-ec5d-4df9-98eb-ae8c8c27486e req-3e6d63cc-8a57-4972-bafe-b2ebce216a29 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Refreshing network info cache for port f2248ec9-0230-4aa8-a016-a30ddf4f45f4 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 927.226190] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339138, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.310912] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.516s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.311119] env[61243]: DEBUG nova.compute.manager [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 927.314019] env[61243]: DEBUG oslo_concurrency.lockutils [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.836s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.314305] env[61243]: DEBUG nova.objects.instance [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lazy-loading 'resources' on Instance uuid c1a0d242-4eae-4c03-8341-840b41341f17 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.326146] env[61243]: DEBUG oslo_vmware.api [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339141, 'name': PowerOffVM_Task, 'duration_secs': 0.289084} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.327098] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 927.328035] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 927.328035] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f17f9daa-f4e8-41b1-872d-597623ae4bbe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.346928] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339137, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.410569] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.410853] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.411101] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleting the datastore file [datastore2] 5c063b4a-3fbb-4222-ab6d-7d27b6b99084 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.411412] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f05244c8-a24b-40d1-8092-fabbbaf9b184 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.420960] env[61243]: DEBUG oslo_vmware.api [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 927.420960] env[61243]: value = "task-1339143" [ 927.420960] env[61243]: _type = "Task" [ 927.420960] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.430807] env[61243]: DEBUG oslo_vmware.api [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339143, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.726435] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339138, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.820717] env[61243]: DEBUG nova.objects.instance [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lazy-loading 'numa_topology' on Instance uuid c1a0d242-4eae-4c03-8341-840b41341f17 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.823907] env[61243]: DEBUG nova.compute.utils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 927.826127] env[61243]: DEBUG nova.compute.manager [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 927.826327] env[61243]: DEBUG nova.network.neutron [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 927.846833] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339137, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.899035] env[61243]: DEBUG nova.policy [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b32f4788825f4a8e82e028cc10347746', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '683c027e4adc4cc4a6b8747a8c8c0de9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 927.933575] env[61243]: DEBUG oslo_vmware.api [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339143, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.324361} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.933864] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.934070] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 927.934277] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 927.934512] env[61243]: INFO nova.compute.manager [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Took 1.15 seconds to destroy the instance on the hypervisor. [ 927.934773] env[61243]: DEBUG oslo.service.loopingcall [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.934976] env[61243]: DEBUG nova.compute.manager [-] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 927.935095] env[61243]: DEBUG nova.network.neutron [-] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 927.956521] env[61243]: DEBUG nova.network.neutron [req-29fe67de-ec5d-4df9-98eb-ae8c8c27486e req-3e6d63cc-8a57-4972-bafe-b2ebce216a29 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Updated VIF entry in instance network info cache for port f2248ec9-0230-4aa8-a016-a30ddf4f45f4. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 927.956920] env[61243]: DEBUG nova.network.neutron [req-29fe67de-ec5d-4df9-98eb-ae8c8c27486e req-3e6d63cc-8a57-4972-bafe-b2ebce216a29 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Updating instance_info_cache with network_info: [{"id": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "address": "fa:16:3e:b0:f1:66", "network": {"id": "49be247d-97e7-4085-8272-4ccf41767d10", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1335943921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "399030c8863346ea97f6da669ff3868b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapf2248ec9-02", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.229657] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339138, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.327064] env[61243]: DEBUG nova.objects.base [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 928.330020] env[61243]: DEBUG nova.compute.manager [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 928.351690] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339137, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.460103] env[61243]: DEBUG oslo_concurrency.lockutils [req-29fe67de-ec5d-4df9-98eb-ae8c8c27486e req-3e6d63cc-8a57-4972-bafe-b2ebce216a29 service nova] Releasing lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.609020] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc4c0ce-b265-4128-8e49-2a69ac0701d1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.616806] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dae94f9-53fd-43c5-8420-66c24ec4eea4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.654913] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "86179fb0-99df-4b10-a815-c19168e9521e" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.656104] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80dcab5-4ac0-4ac0-b71f-285e01e4a6cd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.665718] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b5ae0c6-0edf-49cc-a23e-a6de73c2ab87 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.687261] env[61243]: DEBUG nova.compute.provider_tree [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.692271] env[61243]: DEBUG nova.network.neutron [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Successfully created port: 2333ab37-7807-45d2-a74e-05404217df6d {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 928.727827] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339138, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.850586] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339137, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.107095] env[61243]: DEBUG nova.network.neutron [-] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.148778] env[61243]: DEBUG nova.compute.manager [req-a0d2a0ec-b51a-4b8f-a2f3-245a86853225 req-43cafb96-8fe2-4529-8d98-d19822840963 service nova] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Received event network-vif-deleted-133a7a1c-7381-4bac-8300-4bee50500e62 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.196644] env[61243]: DEBUG nova.scheduler.client.report [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 929.228921] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339138, 'name': ReconfigVM_Task, 'duration_secs': 4.235212} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.229249] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Reconfigured VM instance instance-00000050 to attach disk [datastore1] e9a1c8af-7129-4366-bda1-1c3cf40eeab1/e9a1c8af-7129-4366-bda1-1c3cf40eeab1.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 929.230192] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-735e400e-1a67-4feb-a9ed-4f835ed76ad6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.239474] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 929.239474] env[61243]: value = "task-1339144" [ 929.239474] env[61243]: _type = "Task" [ 929.239474] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.250660] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339144, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.346784] env[61243]: DEBUG nova.compute.manager [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 929.354513] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339137, 'name': MoveVirtualDisk_Task, 'duration_secs': 4.554951} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.354513] env[61243]: INFO nova.virt.vmwareapi.ds_util [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_cbdd30f6-3502-48db-be2c-d4b11e96396c/OSTACK_IMG_cbdd30f6-3502-48db-be2c-d4b11e96396c.vmdk to [datastore1] devstack-image-cache_base/2b9c3b0a-5f89-462e-99b2-0c88500ad15b/2b9c3b0a-5f89-462e-99b2-0c88500ad15b.vmdk. [ 929.354826] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Cleaning up location [datastore1] OSTACK_IMG_cbdd30f6-3502-48db-be2c-d4b11e96396c {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 929.354826] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_cbdd30f6-3502-48db-be2c-d4b11e96396c {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 929.355278] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-19cd83fe-aaa7-4276-b6c3-29609cff68a1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.364402] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 929.364402] env[61243]: value = "task-1339145" [ 929.364402] env[61243]: _type = "Task" [ 929.364402] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.379252] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339145, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.393241] env[61243]: DEBUG nova.virt.hardware [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 929.393486] env[61243]: DEBUG nova.virt.hardware [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 929.393486] env[61243]: DEBUG nova.virt.hardware [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 929.393655] env[61243]: DEBUG nova.virt.hardware [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 929.393910] env[61243]: DEBUG nova.virt.hardware [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 929.394070] env[61243]: DEBUG nova.virt.hardware [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 929.394301] env[61243]: DEBUG nova.virt.hardware [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 929.394473] env[61243]: DEBUG nova.virt.hardware [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 929.394652] env[61243]: DEBUG nova.virt.hardware [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 929.394842] env[61243]: DEBUG nova.virt.hardware [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 929.395053] env[61243]: DEBUG nova.virt.hardware [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 929.396022] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e254317-926b-4947-9064-4f51e5dde802 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.405892] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823b67d7-1f94-40bc-9103-998d3c81ef9c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.522576] env[61243]: DEBUG oslo_vmware.rw_handles [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52edd12d-ba48-558c-a422-062c68f15f84/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 929.523503] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-851c4c92-b68a-43ef-b5c2-4a388dc50ff5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.530897] env[61243]: DEBUG oslo_vmware.rw_handles [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52edd12d-ba48-558c-a422-062c68f15f84/disk-0.vmdk is in state: ready. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 929.532057] env[61243]: ERROR oslo_vmware.rw_handles [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52edd12d-ba48-558c-a422-062c68f15f84/disk-0.vmdk due to incomplete transfer. [ 929.532268] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-4c6b5b02-3db1-4132-adb3-293c37fe0640 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.541081] env[61243]: DEBUG oslo_vmware.rw_handles [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52edd12d-ba48-558c-a422-062c68f15f84/disk-0.vmdk. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 929.541368] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Uploaded image f26780ef-1234-46b6-a400-0b3f1542aed9 to the Glance image server {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 929.543809] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Destroying the VM {{(pid=61243) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 929.544249] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-0377cf1f-94a2-4376-a17c-13b5a2c1771a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.552212] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 929.552212] env[61243]: value = "task-1339146" [ 929.552212] env[61243]: _type = "Task" [ 929.552212] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.562173] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339146, 'name': Destroy_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.608771] env[61243]: INFO nova.compute.manager [-] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Took 1.67 seconds to deallocate network for instance. [ 929.703358] env[61243]: DEBUG oslo_concurrency.lockutils [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.389s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.705909] env[61243]: DEBUG oslo_concurrency.lockutils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.778s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.707632] env[61243]: INFO nova.compute.claims [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 929.750317] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339144, 'name': Rename_Task, 'duration_secs': 0.258685} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.754020] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 929.754020] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6bc070f9-ac0c-4530-bfb2-23ea149988d0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.760927] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 929.760927] env[61243]: value = "task-1339147" [ 929.760927] env[61243]: _type = "Task" [ 929.760927] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.771210] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339147, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.876570] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339145, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.063997} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.877371] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 929.877371] env[61243]: DEBUG oslo_concurrency.lockutils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2b9c3b0a-5f89-462e-99b2-0c88500ad15b/2b9c3b0a-5f89-462e-99b2-0c88500ad15b.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.877800] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2b9c3b0a-5f89-462e-99b2-0c88500ad15b/2b9c3b0a-5f89-462e-99b2-0c88500ad15b.vmdk to [datastore1] 09587d68-f7fd-4648-bd5c-e53b1b6464e8/09587d68-f7fd-4648-bd5c-e53b1b6464e8.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 929.878144] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-903d80f7-7361-4d68-a05e-81e8d2a74b89 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.885047] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 929.885047] env[61243]: value = "task-1339148" [ 929.885047] env[61243]: _type = "Task" [ 929.885047] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.894322] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339148, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.064284] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339146, 'name': Destroy_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.116619] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.222389] env[61243]: DEBUG oslo_concurrency.lockutils [None req-032f0085-cf79-42b6-bc4c-6e8179dea5c3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "c1a0d242-4eae-4c03-8341-840b41341f17" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 31.034s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.222389] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "c1a0d242-4eae-4c03-8341-840b41341f17" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 9.771s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.222389] env[61243]: INFO nova.compute.manager [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Unshelving [ 930.274994] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339147, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.398842] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339148, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.568696] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339146, 'name': Destroy_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.746463] env[61243]: DEBUG nova.network.neutron [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Successfully updated port: 2333ab37-7807-45d2-a74e-05404217df6d {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 930.776351] env[61243]: DEBUG oslo_vmware.api [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339147, 'name': PowerOnVM_Task, 'duration_secs': 1.005356} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.779345] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 930.779918] env[61243]: DEBUG nova.compute.manager [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 930.782207] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40598542-1ad7-45a4-bb59-81e8e1327b90 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.902025] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339148, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.018283] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ecc8fe-4eb3-4ba7-8659-7fa6812798ab {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.032357] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f43a857e-fc20-4914-8cd8-04429b0a69e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.075419] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686ff090-c45f-49e8-9a33-e55ccb15e9d5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.091475] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2602baa0-5927-45d2-ac49-e3f47579fb6c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.098034] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339146, 'name': Destroy_Task, 'duration_secs': 1.427053} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.098034] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Destroyed the VM [ 931.098034] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Deleting Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 931.098034] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-755f5f3e-3f24-4bc9-a2ec-eafc1346ab7b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.109918] env[61243]: DEBUG nova.compute.provider_tree [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.113244] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 931.113244] env[61243]: value = "task-1339149" [ 931.113244] env[61243]: _type = "Task" [ 931.113244] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.123178] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339149, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.189581] env[61243]: DEBUG nova.compute.manager [req-92b0f126-d415-4899-8d4b-ad6adbaa97c6 req-e05fde50-040a-4158-a2d6-e725ff697602 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Received event network-vif-plugged-2333ab37-7807-45d2-a74e-05404217df6d {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.189581] env[61243]: DEBUG oslo_concurrency.lockutils [req-92b0f126-d415-4899-8d4b-ad6adbaa97c6 req-e05fde50-040a-4158-a2d6-e725ff697602 service nova] Acquiring lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.189581] env[61243]: DEBUG oslo_concurrency.lockutils [req-92b0f126-d415-4899-8d4b-ad6adbaa97c6 req-e05fde50-040a-4158-a2d6-e725ff697602 service nova] Lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.189581] env[61243]: DEBUG oslo_concurrency.lockutils [req-92b0f126-d415-4899-8d4b-ad6adbaa97c6 req-e05fde50-040a-4158-a2d6-e725ff697602 service nova] Lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.189581] env[61243]: DEBUG nova.compute.manager [req-92b0f126-d415-4899-8d4b-ad6adbaa97c6 req-e05fde50-040a-4158-a2d6-e725ff697602 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] No waiting events found dispatching network-vif-plugged-2333ab37-7807-45d2-a74e-05404217df6d {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 931.189581] env[61243]: WARNING nova.compute.manager [req-92b0f126-d415-4899-8d4b-ad6adbaa97c6 req-e05fde50-040a-4158-a2d6-e725ff697602 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Received unexpected event network-vif-plugged-2333ab37-7807-45d2-a74e-05404217df6d for instance with vm_state building and task_state spawning. [ 931.189581] env[61243]: DEBUG nova.compute.manager [req-92b0f126-d415-4899-8d4b-ad6adbaa97c6 req-e05fde50-040a-4158-a2d6-e725ff697602 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Received event network-changed-2333ab37-7807-45d2-a74e-05404217df6d {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.189581] env[61243]: DEBUG nova.compute.manager [req-92b0f126-d415-4899-8d4b-ad6adbaa97c6 req-e05fde50-040a-4158-a2d6-e725ff697602 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Refreshing instance network info cache due to event network-changed-2333ab37-7807-45d2-a74e-05404217df6d. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 931.189581] env[61243]: DEBUG oslo_concurrency.lockutils [req-92b0f126-d415-4899-8d4b-ad6adbaa97c6 req-e05fde50-040a-4158-a2d6-e725ff697602 service nova] Acquiring lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.189581] env[61243]: DEBUG oslo_concurrency.lockutils [req-92b0f126-d415-4899-8d4b-ad6adbaa97c6 req-e05fde50-040a-4158-a2d6-e725ff697602 service nova] Acquired lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.189581] env[61243]: DEBUG nova.network.neutron [req-92b0f126-d415-4899-8d4b-ad6adbaa97c6 req-e05fde50-040a-4158-a2d6-e725ff697602 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Refreshing network info cache for port 2333ab37-7807-45d2-a74e-05404217df6d {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 931.238562] env[61243]: DEBUG nova.compute.utils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 931.254029] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.307583] env[61243]: DEBUG oslo_concurrency.lockutils [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.399809] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339148, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.615914] env[61243]: DEBUG nova.scheduler.client.report [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 931.631056] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339149, 'name': RemoveSnapshot_Task} progress is 17%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.743764] env[61243]: INFO nova.virt.block_device [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Booting with volume bc3b5837-80e8-4d8a-8a68-5ec6c3b89880 at /dev/sdb [ 931.753929] env[61243]: DEBUG nova.network.neutron [req-92b0f126-d415-4899-8d4b-ad6adbaa97c6 req-e05fde50-040a-4158-a2d6-e725ff697602 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 931.791450] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2bc596f5-814b-49b6-8080-2525ce1f9e87 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.802538] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87129cb9-38db-4f8c-a7f2-659960c19d49 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.838321] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ec80802-517c-4a7b-83c2-3bcc9339cb49 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.849758] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa90e20-1ccb-4773-bb98-f697e8d8f925 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.866222] env[61243]: DEBUG nova.network.neutron [req-92b0f126-d415-4899-8d4b-ad6adbaa97c6 req-e05fde50-040a-4158-a2d6-e725ff697602 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.884095] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b9e396b-4f1f-44f4-af7f-24a0dae26afb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.895125] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8ca8e3-71db-484e-9e79-7bc82c8db828 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.902892] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339148, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.915859] env[61243]: DEBUG nova.virt.block_device [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Updating existing volume attachment record: 4fe790a5-0885-43c4-8f80-f814824103e5 {{(pid=61243) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 932.120740] env[61243]: DEBUG oslo_concurrency.lockutils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.415s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.126177] env[61243]: DEBUG nova.compute.manager [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 932.126177] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.551s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.126177] env[61243]: DEBUG nova.objects.instance [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Lazy-loading 'resources' on Instance uuid 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.137899] env[61243]: DEBUG oslo_vmware.api [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339149, 'name': RemoveSnapshot_Task, 'duration_secs': 0.621514} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.139033] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Deleted Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 932.139281] env[61243]: INFO nova.compute.manager [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Took 15.33 seconds to snapshot the instance on the hypervisor. [ 932.368555] env[61243]: DEBUG oslo_concurrency.lockutils [req-92b0f126-d415-4899-8d4b-ad6adbaa97c6 req-e05fde50-040a-4158-a2d6-e725ff697602 service nova] Releasing lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.369100] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.369100] env[61243]: DEBUG nova.network.neutron [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 932.398754] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339148, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.634528] env[61243]: DEBUG nova.compute.utils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 932.637170] env[61243]: DEBUG nova.compute.manager [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 932.637170] env[61243]: DEBUG nova.network.neutron [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 932.674083] env[61243]: DEBUG oslo_concurrency.lockutils [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "e9a1c8af-7129-4366-bda1-1c3cf40eeab1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.674363] env[61243]: DEBUG oslo_concurrency.lockutils [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "e9a1c8af-7129-4366-bda1-1c3cf40eeab1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.674577] env[61243]: DEBUG oslo_concurrency.lockutils [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "e9a1c8af-7129-4366-bda1-1c3cf40eeab1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.674828] env[61243]: DEBUG oslo_concurrency.lockutils [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "e9a1c8af-7129-4366-bda1-1c3cf40eeab1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.674935] env[61243]: DEBUG oslo_concurrency.lockutils [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "e9a1c8af-7129-4366-bda1-1c3cf40eeab1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.677550] env[61243]: INFO nova.compute.manager [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Terminating instance [ 932.679420] env[61243]: DEBUG nova.compute.manager [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 932.679648] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 932.680508] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d69b85-0940-4fe7-ac89-82df5d08b32b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.695397] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 932.697953] env[61243]: DEBUG nova.compute.manager [None req-31ebfd3f-4297-4d57-ba8d-560aafe2402f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Found 2 images (rotation: 2) {{(pid=61243) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 932.698937] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f7f6f18-a0d5-4380-94ac-39026c957ae9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.709236] env[61243]: DEBUG nova.policy [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b9748805e2c48e4a8fd0354e916aa0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '324ed556b6f5419fa0376a7cfc508827', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 932.712350] env[61243]: DEBUG oslo_vmware.api [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 932.712350] env[61243]: value = "task-1339153" [ 932.712350] env[61243]: _type = "Task" [ 932.712350] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.725695] env[61243]: DEBUG oslo_vmware.api [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339153, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.901331] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339148, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.573808} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.901755] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2b9c3b0a-5f89-462e-99b2-0c88500ad15b/2b9c3b0a-5f89-462e-99b2-0c88500ad15b.vmdk to [datastore1] 09587d68-f7fd-4648-bd5c-e53b1b6464e8/09587d68-f7fd-4648-bd5c-e53b1b6464e8.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 932.902763] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20b5469-50f6-4377-9579-967c4db574e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.930870] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 09587d68-f7fd-4648-bd5c-e53b1b6464e8/09587d68-f7fd-4648-bd5c-e53b1b6464e8.vmdk or device None with type streamOptimized {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 932.930870] env[61243]: DEBUG nova.network.neutron [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 932.935512] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b03486e8-ccbf-40dc-a3dc-b1d3ca938bf4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.960202] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 932.960202] env[61243]: value = "task-1339154" [ 932.960202] env[61243]: _type = "Task" [ 932.960202] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.971090] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339154, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.002068] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d03f013-7141-4beb-8447-3fe5e648e192 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.012337] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdfa1fa6-0d40-4ea9-8189-a08f28379e45 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.048570] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb84501-7b9a-4b95-b88b-7d1ce308bf23 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.063852] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de6bc035-54c4-4dda-89a0-861eb6072d6e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.083809] env[61243]: DEBUG nova.compute.provider_tree [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.141539] env[61243]: DEBUG nova.compute.manager [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 933.224690] env[61243]: DEBUG oslo_vmware.api [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339153, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.292419] env[61243]: DEBUG nova.network.neutron [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Updating instance_info_cache with network_info: [{"id": "2333ab37-7807-45d2-a74e-05404217df6d", "address": "fa:16:3e:ea:93:91", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2333ab37-78", "ovs_interfaceid": "2333ab37-7807-45d2-a74e-05404217df6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.300466] env[61243]: DEBUG nova.network.neutron [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Successfully created port: 67462132-75b6-4121-9a0d-e65fb2a726d2 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 933.457785] env[61243]: DEBUG oslo_concurrency.lockutils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "7ce043e6-abdf-4708-80aa-221870375c23" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.458096] env[61243]: DEBUG oslo_concurrency.lockutils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "7ce043e6-abdf-4708-80aa-221870375c23" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.469294] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339154, 'name': ReconfigVM_Task, 'duration_secs': 0.298806} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.469600] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 09587d68-f7fd-4648-bd5c-e53b1b6464e8/09587d68-f7fd-4648-bd5c-e53b1b6464e8.vmdk or device None with type streamOptimized {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.470285] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e78be504-750c-4124-b85d-5ceaca4d8754 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.478626] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 933.478626] env[61243]: value = "task-1339155" [ 933.478626] env[61243]: _type = "Task" [ 933.478626] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.487958] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339155, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.563376] env[61243]: DEBUG nova.compute.manager [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 933.564070] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c511868-1c98-47f1-895a-79c1f516e6b4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.591198] env[61243]: DEBUG nova.scheduler.client.report [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.724729] env[61243]: DEBUG oslo_vmware.api [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339153, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.800019] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.800019] env[61243]: DEBUG nova.compute.manager [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Instance network_info: |[{"id": "2333ab37-7807-45d2-a74e-05404217df6d", "address": "fa:16:3e:ea:93:91", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2333ab37-78", "ovs_interfaceid": "2333ab37-7807-45d2-a74e-05404217df6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 933.800019] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ea:93:91', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '93c5b7ce-4c84-40bc-884c-b2453e0eee69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2333ab37-7807-45d2-a74e-05404217df6d', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 933.805964] env[61243]: DEBUG oslo.service.loopingcall [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.806433] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 933.806818] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-96da79ed-468b-43a5-8145-9c23196b7ce2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.827774] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 933.827774] env[61243]: value = "task-1339156" [ 933.827774] env[61243]: _type = "Task" [ 933.827774] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.836262] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339156, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.961098] env[61243]: DEBUG nova.compute.manager [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 933.989301] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339155, 'name': Rename_Task, 'duration_secs': 0.149941} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.989592] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 933.989849] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2110df0-8b4b-4634-9e1a-7230ccce1a11 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.996585] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 933.996585] env[61243]: value = "task-1339157" [ 933.996585] env[61243]: _type = "Task" [ 933.996585] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.007913] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339157, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.076017] env[61243]: INFO nova.compute.manager [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] instance snapshotting [ 934.076690] env[61243]: DEBUG nova.objects.instance [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lazy-loading 'flavor' on Instance uuid 316572e0-c007-42cb-aaf0-3a8cfcaf24aa {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.096662] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.972s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.099652] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.739s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.100299] env[61243]: DEBUG nova.objects.instance [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lazy-loading 'resources' on Instance uuid 894eb798-6c9e-47cb-8eb5-4610fe184bc5 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.129141] env[61243]: INFO nova.scheduler.client.report [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Deleted allocations for instance 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f [ 934.151490] env[61243]: DEBUG nova.compute.manager [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 934.186032] env[61243]: DEBUG nova.virt.hardware [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 934.186207] env[61243]: DEBUG nova.virt.hardware [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 934.186318] env[61243]: DEBUG nova.virt.hardware [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 934.186511] env[61243]: DEBUG nova.virt.hardware [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 934.186664] env[61243]: DEBUG nova.virt.hardware [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 934.186829] env[61243]: DEBUG nova.virt.hardware [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 934.187061] env[61243]: DEBUG nova.virt.hardware [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 934.188558] env[61243]: DEBUG nova.virt.hardware [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 934.188558] env[61243]: DEBUG nova.virt.hardware [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 934.188558] env[61243]: DEBUG nova.virt.hardware [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 934.188558] env[61243]: DEBUG nova.virt.hardware [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 934.188817] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10895f47-23e7-4cad-a741-6525ea8c4129 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.198555] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc27452-cf13-47b0-89d8-64298ad24e3e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.224778] env[61243]: DEBUG oslo_vmware.api [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339153, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.342879] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339156, 'name': CreateVM_Task, 'duration_secs': 0.366494} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.343145] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 934.344310] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.344529] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.345443] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 934.345813] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34cd573e-3e8f-4662-948d-f0e4ec1ffa98 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.352816] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 934.352816] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5258943c-3f94-7b92-9b3d-01dce30c8a15" [ 934.352816] env[61243]: _type = "Task" [ 934.352816] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.364820] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5258943c-3f94-7b92-9b3d-01dce30c8a15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.484512] env[61243]: DEBUG oslo_concurrency.lockutils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.511444] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339157, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.585088] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb581d0-410e-40f8-bdae-8658bdec9610 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.610420] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2fd7243-8885-4d4a-bd37-f2a90e673c80 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.638543] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9324387e-46e4-4c39-8154-01f02bf842e9 tempest-ServersTestManualDisk-76506012 tempest-ServersTestManualDisk-76506012-project-member] Lock "8df486c4-d1ce-47b0-81a8-ad9e83eaac7f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.966s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.734710] env[61243]: DEBUG oslo_vmware.api [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339153, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.833929] env[61243]: DEBUG nova.compute.manager [req-466fb4dc-cb93-44de-b1e6-b13b422f8623 req-95ef44e1-f643-44ba-b4bf-386d7c637482 service nova] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Received event network-vif-plugged-67462132-75b6-4121-9a0d-e65fb2a726d2 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.834327] env[61243]: DEBUG oslo_concurrency.lockutils [req-466fb4dc-cb93-44de-b1e6-b13b422f8623 req-95ef44e1-f643-44ba-b4bf-386d7c637482 service nova] Acquiring lock "f77c2610-fa49-487c-9fb4-3b112f800daf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.835021] env[61243]: DEBUG oslo_concurrency.lockutils [req-466fb4dc-cb93-44de-b1e6-b13b422f8623 req-95ef44e1-f643-44ba-b4bf-386d7c637482 service nova] Lock "f77c2610-fa49-487c-9fb4-3b112f800daf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.835021] env[61243]: DEBUG oslo_concurrency.lockutils [req-466fb4dc-cb93-44de-b1e6-b13b422f8623 req-95ef44e1-f643-44ba-b4bf-386d7c637482 service nova] Lock "f77c2610-fa49-487c-9fb4-3b112f800daf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.836367] env[61243]: DEBUG nova.compute.manager [req-466fb4dc-cb93-44de-b1e6-b13b422f8623 req-95ef44e1-f643-44ba-b4bf-386d7c637482 service nova] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] No waiting events found dispatching network-vif-plugged-67462132-75b6-4121-9a0d-e65fb2a726d2 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 934.836367] env[61243]: WARNING nova.compute.manager [req-466fb4dc-cb93-44de-b1e6-b13b422f8623 req-95ef44e1-f643-44ba-b4bf-386d7c637482 service nova] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Received unexpected event network-vif-plugged-67462132-75b6-4121-9a0d-e65fb2a726d2 for instance with vm_state building and task_state spawning. [ 934.870372] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5258943c-3f94-7b92-9b3d-01dce30c8a15, 'name': SearchDatastore_Task, 'duration_secs': 0.012639} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.870460] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.870718] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 934.870999] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.871255] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.871505] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 934.871853] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ba8091fa-e9a2-4f60-b5df-ad967b923a68 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.875315] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc091008-7b9d-4559-9ed9-afdb866673e3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.884960] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d60ccff-1ab5-4c7a-a916-52d69ab5f788 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.889386] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 934.889573] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 934.890741] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e0f604c-e624-4d87-a882-848e671de30a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.918759] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7938a8-9bf8-4e87-84a9-8e35a603e7d8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.923183] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 934.923183] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5255e3a0-ed80-0bf6-5940-eac88cac4714" [ 934.923183] env[61243]: _type = "Task" [ 934.923183] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.931427] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09f3245-5a0d-4476-b6ef-5c5d999c066a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.939349] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5255e3a0-ed80-0bf6-5940-eac88cac4714, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.950721] env[61243]: DEBUG nova.compute.provider_tree [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.953011] env[61243]: DEBUG nova.network.neutron [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Successfully updated port: 67462132-75b6-4121-9a0d-e65fb2a726d2 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 935.008755] env[61243]: DEBUG oslo_vmware.api [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339157, 'name': PowerOnVM_Task, 'duration_secs': 0.614899} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.009046] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 935.009259] env[61243]: INFO nova.compute.manager [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Took 17.00 seconds to spawn the instance on the hypervisor. [ 935.009462] env[61243]: DEBUG nova.compute.manager [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 935.010308] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85cf5fd8-134d-44f8-a3f6-4fda1f9964a0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.126824] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Creating Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 935.127178] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cf10308e-f647-4c84-8928-541954cbf3e5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.137302] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 935.137302] env[61243]: value = "task-1339159" [ 935.137302] env[61243]: _type = "Task" [ 935.137302] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.149783] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339159, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.230310] env[61243]: DEBUG oslo_vmware.api [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339153, 'name': PowerOffVM_Task, 'duration_secs': 2.429535} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.230594] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.230814] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 935.231136] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b1511403-5ba6-40f9-b021-73f19bc584c9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.309096] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 935.309375] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 935.309582] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleting the datastore file [datastore1] e9a1c8af-7129-4366-bda1-1c3cf40eeab1 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.309866] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28370a05-518f-417d-9e6c-38a3bd8ed957 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.318668] env[61243]: DEBUG oslo_vmware.api [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 935.318668] env[61243]: value = "task-1339161" [ 935.318668] env[61243]: _type = "Task" [ 935.318668] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.328774] env[61243]: DEBUG oslo_vmware.api [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339161, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.435398] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5255e3a0-ed80-0bf6-5940-eac88cac4714, 'name': SearchDatastore_Task, 'duration_secs': 0.016861} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.436230] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecfcf0b9-c7d7-4cfc-8321-8f89e094d1b9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.442635] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 935.442635] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52acf322-247d-b107-dd06-d83e3ca3f3bb" [ 935.442635] env[61243]: _type = "Task" [ 935.442635] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.451268] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52acf322-247d-b107-dd06-d83e3ca3f3bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.455312] env[61243]: DEBUG nova.scheduler.client.report [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.459885] env[61243]: DEBUG oslo_concurrency.lockutils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "refresh_cache-f77c2610-fa49-487c-9fb4-3b112f800daf" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.459885] env[61243]: DEBUG oslo_concurrency.lockutils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "refresh_cache-f77c2610-fa49-487c-9fb4-3b112f800daf" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.460020] env[61243]: DEBUG nova.network.neutron [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 935.527503] env[61243]: INFO nova.compute.manager [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Took 34.58 seconds to build instance. [ 935.650546] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339159, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.829656] env[61243]: DEBUG oslo_vmware.api [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339161, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209307} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.830090] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.830247] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.830353] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.830535] env[61243]: INFO nova.compute.manager [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Took 3.15 seconds to destroy the instance on the hypervisor. [ 935.830786] env[61243]: DEBUG oslo.service.loopingcall [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.831016] env[61243]: DEBUG nova.compute.manager [-] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 935.831141] env[61243]: DEBUG nova.network.neutron [-] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 935.956298] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52acf322-247d-b107-dd06-d83e3ca3f3bb, 'name': SearchDatastore_Task, 'duration_secs': 0.011464} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.956298] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.956298] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] b0b5b215-d7ff-4960-9d39-cf792d8d4656/b0b5b215-d7ff-4960-9d39-cf792d8d4656.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 935.956298] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fa7f342e-e3e2-480e-a7f6-f74dbcda9eef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.965166] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.865s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.968895] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 935.968895] env[61243]: value = "task-1339162" [ 935.968895] env[61243]: _type = "Task" [ 935.968895] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.969845] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.749s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.970298] env[61243]: DEBUG nova.objects.instance [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lazy-loading 'resources' on Instance uuid ebe8cc8e-cc35-4ec4-bed7-d55df72a2991 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.989104] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339162, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.999716] env[61243]: INFO nova.scheduler.client.report [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Deleted allocations for instance 894eb798-6c9e-47cb-8eb5-4610fe184bc5 [ 936.024431] env[61243]: DEBUG nova.network.neutron [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 936.030107] env[61243]: DEBUG oslo_concurrency.lockutils [None req-72436ce3-f3b4-422d-bc46-08932738c3ec tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "09587d68-f7fd-4648-bd5c-e53b1b6464e8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.092s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.061834] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "09587d68-f7fd-4648-bd5c-e53b1b6464e8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.062559] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "09587d68-f7fd-4648-bd5c-e53b1b6464e8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.063037] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "09587d68-f7fd-4648-bd5c-e53b1b6464e8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.063386] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "09587d68-f7fd-4648-bd5c-e53b1b6464e8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.063811] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "09587d68-f7fd-4648-bd5c-e53b1b6464e8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.069277] env[61243]: INFO nova.compute.manager [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Terminating instance [ 936.073528] env[61243]: DEBUG nova.compute.manager [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 936.074305] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 936.077303] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd41bd9-6d84-4e39-8a90-a98d13716366 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.087147] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 936.087353] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a518af01-b47a-4a39-bd43-c47dc8eacc8a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.103024] env[61243]: DEBUG oslo_vmware.api [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 936.103024] env[61243]: value = "task-1339163" [ 936.103024] env[61243]: _type = "Task" [ 936.103024] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.117255] env[61243]: DEBUG oslo_vmware.api [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339163, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.152740] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339159, 'name': CreateSnapshot_Task, 'duration_secs': 0.97116} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.153210] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Created Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 936.154455] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29523560-775f-40a5-ae1a-0442786d6cc1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.233258] env[61243]: DEBUG nova.network.neutron [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Updating instance_info_cache with network_info: [{"id": "67462132-75b6-4121-9a0d-e65fb2a726d2", "address": "fa:16:3e:87:72:30", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67462132-75", "ovs_interfaceid": "67462132-75b6-4121-9a0d-e65fb2a726d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.489067] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339162, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518653} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.489385] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] b0b5b215-d7ff-4960-9d39-cf792d8d4656/b0b5b215-d7ff-4960-9d39-cf792d8d4656.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 936.490040] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 936.490040] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-69c06b9d-781d-4b49-ab94-55759ec147f5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.502183] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 936.502183] env[61243]: value = "task-1339164" [ 936.502183] env[61243]: _type = "Task" [ 936.502183] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.511267] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f7f10dd1-b4d4-4789-8811-aa6055999a18 tempest-ServerRescueTestJSON-705512608 tempest-ServerRescueTestJSON-705512608-project-member] Lock "894eb798-6c9e-47cb-8eb5-4610fe184bc5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.742s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.519169] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339164, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.616212] env[61243]: DEBUG oslo_vmware.api [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339163, 'name': PowerOffVM_Task, 'duration_secs': 0.193128} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.619069] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 936.619343] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 936.619738] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a329857f-024c-441e-82c5-1c3926ad9d13 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.680184] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Creating linked-clone VM from snapshot {{(pid=61243) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 936.680542] env[61243]: DEBUG nova.network.neutron [-] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.685025] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1fe3d33b-052f-49bb-b430-c79fc8f31981 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.693380] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 936.693530] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 936.693770] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Deleting the datastore file [datastore1] 09587d68-f7fd-4648-bd5c-e53b1b6464e8 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 936.694995] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5e1129ae-314e-47d5-bddf-213e4dbdb006 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.698406] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 936.698406] env[61243]: value = "task-1339166" [ 936.698406] env[61243]: _type = "Task" [ 936.698406] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.707111] env[61243]: DEBUG oslo_vmware.api [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 936.707111] env[61243]: value = "task-1339167" [ 936.707111] env[61243]: _type = "Task" [ 936.707111] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.717983] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339166, 'name': CloneVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.727156] env[61243]: DEBUG oslo_vmware.api [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339167, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.737106] env[61243]: DEBUG oslo_concurrency.lockutils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "refresh_cache-f77c2610-fa49-487c-9fb4-3b112f800daf" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.737775] env[61243]: DEBUG nova.compute.manager [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Instance network_info: |[{"id": "67462132-75b6-4121-9a0d-e65fb2a726d2", "address": "fa:16:3e:87:72:30", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67462132-75", "ovs_interfaceid": "67462132-75b6-4121-9a0d-e65fb2a726d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 936.737910] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:72:30', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ff3ecd2f-0b10-4faf-a512-fd7a20c28df1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '67462132-75b6-4121-9a0d-e65fb2a726d2', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 936.745965] env[61243]: DEBUG oslo.service.loopingcall [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.747540] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 936.748367] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb3488af-16db-4ec8-afac-d9adf69706ff {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.751048] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ac2ff63b-a5f3-4994-bc03-2727b21daf5e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.774722] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c27f3eb-f6fb-49f5-a564-8047757f5d52 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.777632] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 936.777632] env[61243]: value = "task-1339168" [ 936.777632] env[61243]: _type = "Task" [ 936.777632] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.812927] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed2610d0-9a76-43dc-aba6-5f35e6f77943 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.815835] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339168, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.822219] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5573897d-bd84-4f6a-b8cb-85ba127634d7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.843762] env[61243]: DEBUG nova.compute.provider_tree [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.871464] env[61243]: DEBUG nova.compute.manager [req-abddac24-b96d-4682-ac20-72e258e8f3f7 req-f70a77c8-161e-4804-a7f9-25ba92732b67 service nova] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Received event network-changed-67462132-75b6-4121-9a0d-e65fb2a726d2 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.871899] env[61243]: DEBUG nova.compute.manager [req-abddac24-b96d-4682-ac20-72e258e8f3f7 req-f70a77c8-161e-4804-a7f9-25ba92732b67 service nova] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Refreshing instance network info cache due to event network-changed-67462132-75b6-4121-9a0d-e65fb2a726d2. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 936.872086] env[61243]: DEBUG oslo_concurrency.lockutils [req-abddac24-b96d-4682-ac20-72e258e8f3f7 req-f70a77c8-161e-4804-a7f9-25ba92732b67 service nova] Acquiring lock "refresh_cache-f77c2610-fa49-487c-9fb4-3b112f800daf" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.872169] env[61243]: DEBUG oslo_concurrency.lockutils [req-abddac24-b96d-4682-ac20-72e258e8f3f7 req-f70a77c8-161e-4804-a7f9-25ba92732b67 service nova] Acquired lock "refresh_cache-f77c2610-fa49-487c-9fb4-3b112f800daf" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.872443] env[61243]: DEBUG nova.network.neutron [req-abddac24-b96d-4682-ac20-72e258e8f3f7 req-f70a77c8-161e-4804-a7f9-25ba92732b67 service nova] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Refreshing network info cache for port 67462132-75b6-4121-9a0d-e65fb2a726d2 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 937.010839] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339164, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.251528} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.011164] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 937.011978] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17bfec31-2165-4891-a405-be574e266c8e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.044471] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] b0b5b215-d7ff-4960-9d39-cf792d8d4656/b0b5b215-d7ff-4960-9d39-cf792d8d4656.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 937.045584] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b92e1142-43f9-444a-b532-c41b44ca93cb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.070428] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 937.070428] env[61243]: value = "task-1339169" [ 937.070428] env[61243]: _type = "Task" [ 937.070428] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.085146] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339169, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.189933] env[61243]: INFO nova.compute.manager [-] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Took 1.36 seconds to deallocate network for instance. [ 937.210771] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339166, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.223112] env[61243]: DEBUG oslo_vmware.api [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339167, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.230604} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.223408] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 937.223609] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 937.223985] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 937.224144] env[61243]: INFO nova.compute.manager [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Took 1.15 seconds to destroy the instance on the hypervisor. [ 937.224777] env[61243]: DEBUG oslo.service.loopingcall [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.225014] env[61243]: DEBUG nova.compute.manager [-] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 937.225112] env[61243]: DEBUG nova.network.neutron [-] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 937.293724] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339168, 'name': CreateVM_Task, 'duration_secs': 0.429237} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.293925] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 937.294659] env[61243]: DEBUG oslo_concurrency.lockutils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.294859] env[61243]: DEBUG oslo_concurrency.lockutils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.295290] env[61243]: DEBUG oslo_concurrency.lockutils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 937.295582] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51cd7466-5901-446c-aaf7-40b33fb8096e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.303143] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 937.303143] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5294f2cb-1332-f9c9-98b5-051e35c3d023" [ 937.303143] env[61243]: _type = "Task" [ 937.303143] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.312628] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5294f2cb-1332-f9c9-98b5-051e35c3d023, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.347931] env[61243]: DEBUG nova.scheduler.client.report [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.537613] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.583534] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339169, 'name': ReconfigVM_Task, 'duration_secs': 0.324083} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.583534] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Reconfigured VM instance instance-00000055 to attach disk [datastore2] b0b5b215-d7ff-4960-9d39-cf792d8d4656/b0b5b215-d7ff-4960-9d39-cf792d8d4656.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.584456] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a6f1ce90-0f1f-436e-842e-0188fd422864 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.592950] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 937.592950] env[61243]: value = "task-1339170" [ 937.592950] env[61243]: _type = "Task" [ 937.592950] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.607140] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339170, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.696970] env[61243]: DEBUG oslo_concurrency.lockutils [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.711276] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339166, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.817426] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5294f2cb-1332-f9c9-98b5-051e35c3d023, 'name': SearchDatastore_Task, 'duration_secs': 0.018086} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.817760] env[61243]: DEBUG oslo_concurrency.lockutils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.818156] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 937.818471] env[61243]: DEBUG oslo_concurrency.lockutils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.818658] env[61243]: DEBUG oslo_concurrency.lockutils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.818852] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 937.819179] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9ad96808-e3c4-423c-8714-8aa3767bfd49 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.831901] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 937.832166] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 937.832950] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2c1762b-f86f-47dd-bda1-6b8817783fc6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.839634] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 937.839634] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525ee5bf-8e97-4e7e-5f10-5c0a768a3cb0" [ 937.839634] env[61243]: _type = "Task" [ 937.839634] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.850378] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525ee5bf-8e97-4e7e-5f10-5c0a768a3cb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.856179] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.886s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.858258] env[61243]: DEBUG oslo_concurrency.lockutils [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.065s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.858517] env[61243]: DEBUG nova.objects.instance [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lazy-loading 'resources' on Instance uuid 86179fb0-99df-4b10-a815-c19168e9521e {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.884235] env[61243]: INFO nova.scheduler.client.report [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Deleted allocations for instance ebe8cc8e-cc35-4ec4-bed7-d55df72a2991 [ 938.115418] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339170, 'name': Rename_Task, 'duration_secs': 0.240177} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.115418] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 938.115418] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-34a2bdd3-f3b1-43dd-b963-fe329a71c022 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.124913] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 938.124913] env[61243]: value = "task-1339171" [ 938.124913] env[61243]: _type = "Task" [ 938.124913] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.138841] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339171, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.168726] env[61243]: DEBUG nova.network.neutron [req-abddac24-b96d-4682-ac20-72e258e8f3f7 req-f70a77c8-161e-4804-a7f9-25ba92732b67 service nova] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Updated VIF entry in instance network info cache for port 67462132-75b6-4121-9a0d-e65fb2a726d2. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 938.168926] env[61243]: DEBUG nova.network.neutron [req-abddac24-b96d-4682-ac20-72e258e8f3f7 req-f70a77c8-161e-4804-a7f9-25ba92732b67 service nova] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Updating instance_info_cache with network_info: [{"id": "67462132-75b6-4121-9a0d-e65fb2a726d2", "address": "fa:16:3e:87:72:30", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67462132-75", "ovs_interfaceid": "67462132-75b6-4121-9a0d-e65fb2a726d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.211602] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339166, 'name': CloneVM_Task} progress is 95%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.353468] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]525ee5bf-8e97-4e7e-5f10-5c0a768a3cb0, 'name': SearchDatastore_Task, 'duration_secs': 0.010913} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.353468] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53117ab3-d505-41bd-aebc-fabbcebbac09 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.362037] env[61243]: DEBUG nova.objects.instance [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lazy-loading 'numa_topology' on Instance uuid 86179fb0-99df-4b10-a815-c19168e9521e {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.362774] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 938.362774] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52655d75-3d17-d6e4-1079-de996f42b433" [ 938.362774] env[61243]: _type = "Task" [ 938.362774] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.373074] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52655d75-3d17-d6e4-1079-de996f42b433, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.396863] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4cddbcdd-d1a8-4105-af78-0f0db372f00d tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "ebe8cc8e-cc35-4ec4-bed7-d55df72a2991" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.992s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.591041] env[61243]: DEBUG nova.network.neutron [-] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.637182] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339171, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.671948] env[61243]: DEBUG oslo_concurrency.lockutils [req-abddac24-b96d-4682-ac20-72e258e8f3f7 req-f70a77c8-161e-4804-a7f9-25ba92732b67 service nova] Releasing lock "refresh_cache-f77c2610-fa49-487c-9fb4-3b112f800daf" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.672420] env[61243]: DEBUG nova.compute.manager [req-abddac24-b96d-4682-ac20-72e258e8f3f7 req-f70a77c8-161e-4804-a7f9-25ba92732b67 service nova] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Received event network-vif-deleted-c7f2cf3c-bad1-4fa9-a3e7-be9bf0f8cda3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.711750] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339166, 'name': CloneVM_Task, 'duration_secs': 1.893723} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.712601] env[61243]: INFO nova.virt.vmwareapi.vmops [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Created linked-clone VM from snapshot [ 938.713021] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2af2c4d-050f-4558-9458-0153aab76e6d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.722979] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Uploading image aef980b5-b4fe-49b2-97ec-338458c642cc {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 938.749537] env[61243]: DEBUG oslo_vmware.rw_handles [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 938.749537] env[61243]: value = "vm-285750" [ 938.749537] env[61243]: _type = "VirtualMachine" [ 938.749537] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 938.749859] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3ae31644-fd07-495e-8d5f-a67d31c88a49 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.758933] env[61243]: DEBUG oslo_vmware.rw_handles [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lease: (returnval){ [ 938.758933] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cc6a7d-2f35-8f44-6578-e61716fbfcd1" [ 938.758933] env[61243]: _type = "HttpNfcLease" [ 938.758933] env[61243]: } obtained for exporting VM: (result){ [ 938.758933] env[61243]: value = "vm-285750" [ 938.758933] env[61243]: _type = "VirtualMachine" [ 938.758933] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 938.759306] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the lease: (returnval){ [ 938.759306] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cc6a7d-2f35-8f44-6578-e61716fbfcd1" [ 938.759306] env[61243]: _type = "HttpNfcLease" [ 938.759306] env[61243]: } to be ready. {{(pid=61243) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 938.767689] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 938.767689] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cc6a7d-2f35-8f44-6578-e61716fbfcd1" [ 938.767689] env[61243]: _type = "HttpNfcLease" [ 938.767689] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 938.868588] env[61243]: DEBUG nova.objects.base [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Object Instance<86179fb0-99df-4b10-a815-c19168e9521e> lazy-loaded attributes: resources,numa_topology {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 938.880348] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52655d75-3d17-d6e4-1079-de996f42b433, 'name': SearchDatastore_Task, 'duration_secs': 0.011044} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.880917] env[61243]: DEBUG oslo_concurrency.lockutils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.881786] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] f77c2610-fa49-487c-9fb4-3b112f800daf/f77c2610-fa49-487c-9fb4-3b112f800daf.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 938.881905] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d6f0c1f9-534b-4058-b102-6228d9e7b8e7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.892021] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 938.892021] env[61243]: value = "task-1339173" [ 938.892021] env[61243]: _type = "Task" [ 938.892021] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.902957] env[61243]: DEBUG nova.compute.manager [req-8e9c8220-ee95-4d0f-990b-fa297e03687e req-22fd5ce7-1c0a-4ed0-9d11-829f99656de7 service nova] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Received event network-vif-deleted-99fff1a2-f410-40f3-9e32-fa6bb307a34a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.906904] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339173, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.094761] env[61243]: INFO nova.compute.manager [-] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Took 1.87 seconds to deallocate network for instance. [ 939.103390] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f9607a6-a6e0-4655-82df-12c08819534f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.114823] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-810ac8a3-51da-4225-8f65-4ae1e11a6ced {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.150617] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf2cecd-c9c9-4f80-bb46-e2e1873caf75 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.160222] env[61243]: DEBUG oslo_vmware.api [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339171, 'name': PowerOnVM_Task, 'duration_secs': 0.548389} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.162700] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 939.162941] env[61243]: INFO nova.compute.manager [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Took 9.82 seconds to spawn the instance on the hypervisor. [ 939.163141] env[61243]: DEBUG nova.compute.manager [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 939.163999] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c989681-a6c7-43a5-9472-2f3b4d1b0914 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.167584] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb4e27f2-88a3-49c9-a6b1-87f5f9d224f3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.183663] env[61243]: DEBUG nova.compute.provider_tree [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.269998] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 939.269998] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cc6a7d-2f35-8f44-6578-e61716fbfcd1" [ 939.269998] env[61243]: _type = "HttpNfcLease" [ 939.269998] env[61243]: } is ready. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 939.270431] env[61243]: DEBUG oslo_vmware.rw_handles [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 939.270431] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cc6a7d-2f35-8f44-6578-e61716fbfcd1" [ 939.270431] env[61243]: _type = "HttpNfcLease" [ 939.270431] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 939.271231] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9d5a814-e507-48d4-88f9-eda42c936de8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.280691] env[61243]: DEBUG oslo_vmware.rw_handles [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520d2904-1f16-6c39-c642-66457249f797/disk-0.vmdk from lease info. {{(pid=61243) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 939.280928] env[61243]: DEBUG oslo_vmware.rw_handles [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520d2904-1f16-6c39-c642-66457249f797/disk-0.vmdk for reading. {{(pid=61243) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 939.377152] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c70191e3-b7b3-49a0-a668-b60f1486819b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.410845] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339173, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514401} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.413969] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] f77c2610-fa49-487c-9fb4-3b112f800daf/f77c2610-fa49-487c-9fb4-3b112f800daf.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 939.414344] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 939.414849] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8e89a0c2-7c41-454d-a870-71f2e113db95 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.422915] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 939.422915] env[61243]: value = "task-1339174" [ 939.422915] env[61243]: _type = "Task" [ 939.422915] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.442804] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339174, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.609909] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.687620] env[61243]: DEBUG nova.scheduler.client.report [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.697012] env[61243]: INFO nova.compute.manager [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Took 26.35 seconds to build instance. [ 939.943818] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.944422] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.949552] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339174, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.16549} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.950436] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 939.951373] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd5605d-a3e2-41a5-bdbf-d34b1c744bf3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.976280] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] f77c2610-fa49-487c-9fb4-3b112f800daf/f77c2610-fa49-487c-9fb4-3b112f800daf.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.977442] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c054a7c-88b9-4f2e-98ac-ebdc7306c257 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.001983] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 940.001983] env[61243]: value = "task-1339175" [ 940.001983] env[61243]: _type = "Task" [ 940.001983] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.010140] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339175, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.195926] env[61243]: DEBUG oslo_concurrency.lockutils [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.338s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.199034] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.083s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.199811] env[61243]: DEBUG nova.objects.instance [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lazy-loading 'resources' on Instance uuid 5c063b4a-3fbb-4222-ab6d-7d27b6b99084 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.202067] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b2e040e1-7e03-4622-aa75-18655e6c8645 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.880s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.450757] env[61243]: DEBUG nova.compute.manager [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 940.512246] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339175, 'name': ReconfigVM_Task, 'duration_secs': 0.346137} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.513913] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Reconfigured VM instance instance-00000056 to attach disk [datastore2] f77c2610-fa49-487c-9fb4-3b112f800daf/f77c2610-fa49-487c-9fb4-3b112f800daf.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 940.513913] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65bce883-87dd-44ba-bbf1-9f16526e5cb8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.522052] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 940.522052] env[61243]: value = "task-1339176" [ 940.522052] env[61243]: _type = "Task" [ 940.522052] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.533168] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339176, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.712161] env[61243]: DEBUG oslo_concurrency.lockutils [None req-be53bf24-f5e6-4208-902a-33ec8082e5cc tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "86179fb0-99df-4b10-a815-c19168e9521e" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 37.059s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.713617] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "86179fb0-99df-4b10-a815-c19168e9521e" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 12.059s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.713974] env[61243]: INFO nova.compute.manager [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Unshelving [ 940.898102] env[61243]: DEBUG nova.compute.manager [req-ce6a4419-fc68-42f2-a56e-27622329777d req-34d1a20b-e20f-4df2-8aa1-b689a53e3806 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Received event network-changed-2333ab37-7807-45d2-a74e-05404217df6d {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.898403] env[61243]: DEBUG nova.compute.manager [req-ce6a4419-fc68-42f2-a56e-27622329777d req-34d1a20b-e20f-4df2-8aa1-b689a53e3806 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Refreshing instance network info cache due to event network-changed-2333ab37-7807-45d2-a74e-05404217df6d. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 940.898675] env[61243]: DEBUG oslo_concurrency.lockutils [req-ce6a4419-fc68-42f2-a56e-27622329777d req-34d1a20b-e20f-4df2-8aa1-b689a53e3806 service nova] Acquiring lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.898959] env[61243]: DEBUG oslo_concurrency.lockutils [req-ce6a4419-fc68-42f2-a56e-27622329777d req-34d1a20b-e20f-4df2-8aa1-b689a53e3806 service nova] Acquired lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.899302] env[61243]: DEBUG nova.network.neutron [req-ce6a4419-fc68-42f2-a56e-27622329777d req-34d1a20b-e20f-4df2-8aa1-b689a53e3806 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Refreshing network info cache for port 2333ab37-7807-45d2-a74e-05404217df6d {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 940.938482] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5736159b-b605-4f5d-b87a-0c6c0d594894 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.947231] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb8c27f2-83d4-444b-8395-5266fc3ee34c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.988074] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067348ba-17e4-4200-a22a-29716328f54a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.998031] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-802fa6be-06a0-4209-b7bf-587cbd797a34 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.004672] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.016211] env[61243]: DEBUG nova.compute.provider_tree [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.035754] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339176, 'name': Rename_Task, 'duration_secs': 0.194477} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.036516] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 941.037018] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b193ab16-de6e-47bc-a3ee-0b3607da4e5f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.057806] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 941.057806] env[61243]: value = "task-1339177" [ 941.057806] env[61243]: _type = "Task" [ 941.057806] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.068076] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339177, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.520704] env[61243]: DEBUG nova.scheduler.client.report [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 941.576953] env[61243]: DEBUG oslo_vmware.api [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339177, 'name': PowerOnVM_Task, 'duration_secs': 0.497693} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.577338] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 941.577528] env[61243]: INFO nova.compute.manager [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Took 7.43 seconds to spawn the instance on the hypervisor. [ 941.577692] env[61243]: DEBUG nova.compute.manager [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.578703] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0dfcce-eeab-4a5d-bcf4-7ecff32b4097 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.658525] env[61243]: DEBUG nova.network.neutron [req-ce6a4419-fc68-42f2-a56e-27622329777d req-34d1a20b-e20f-4df2-8aa1-b689a53e3806 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Updated VIF entry in instance network info cache for port 2333ab37-7807-45d2-a74e-05404217df6d. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 941.658905] env[61243]: DEBUG nova.network.neutron [req-ce6a4419-fc68-42f2-a56e-27622329777d req-34d1a20b-e20f-4df2-8aa1-b689a53e3806 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Updating instance_info_cache with network_info: [{"id": "2333ab37-7807-45d2-a74e-05404217df6d", "address": "fa:16:3e:ea:93:91", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2333ab37-78", "ovs_interfaceid": "2333ab37-7807-45d2-a74e-05404217df6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.740538] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.029917] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.831s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.033106] env[61243]: DEBUG oslo_concurrency.lockutils [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 10.726s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.033274] env[61243]: DEBUG nova.objects.instance [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61243) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 942.050562] env[61243]: INFO nova.scheduler.client.report [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleted allocations for instance 5c063b4a-3fbb-4222-ab6d-7d27b6b99084 [ 942.100708] env[61243]: INFO nova.compute.manager [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Took 22.19 seconds to build instance. [ 942.162452] env[61243]: DEBUG oslo_concurrency.lockutils [req-ce6a4419-fc68-42f2-a56e-27622329777d req-34d1a20b-e20f-4df2-8aa1-b689a53e3806 service nova] Releasing lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.557541] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e906bba2-24ea-48f9-a8aa-b3553c647823 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "5c063b4a-3fbb-4222-ab6d-7d27b6b99084" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.777s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.602508] env[61243]: DEBUG oslo_concurrency.lockutils [None req-af4a80f2-aa2d-428d-8341-0e32050ae826 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "f77c2610-fa49-487c-9fb4-3b112f800daf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.712s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.635077] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9617d248-0c39-4b36-b53c-acc48003a150 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "f77c2610-fa49-487c-9fb4-3b112f800daf" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.635269] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9617d248-0c39-4b36-b53c-acc48003a150 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "f77c2610-fa49-487c-9fb4-3b112f800daf" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.635468] env[61243]: DEBUG nova.compute.manager [None req-9617d248-0c39-4b36-b53c-acc48003a150 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 942.636903] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adfd8c08-44cd-44b3-9e03-edd9a0ee2c77 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.644595] env[61243]: DEBUG nova.compute.manager [None req-9617d248-0c39-4b36-b53c-acc48003a150 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61243) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 942.645184] env[61243]: DEBUG nova.objects.instance [None req-9617d248-0c39-4b36-b53c-acc48003a150 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lazy-loading 'flavor' on Instance uuid f77c2610-fa49-487c-9fb4-3b112f800daf {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.043516] env[61243]: DEBUG oslo_concurrency.lockutils [None req-900fd6b9-02e8-4415-a84e-daee917ab12f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.044682] env[61243]: DEBUG oslo_concurrency.lockutils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.561s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.046775] env[61243]: INFO nova.compute.claims [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 943.150577] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9617d248-0c39-4b36-b53c-acc48003a150 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 943.150869] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed9c8bf3-8e65-48cb-b3d2-7297a71f79bb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.159156] env[61243]: DEBUG oslo_vmware.api [None req-9617d248-0c39-4b36-b53c-acc48003a150 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 943.159156] env[61243]: value = "task-1339178" [ 943.159156] env[61243]: _type = "Task" [ 943.159156] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.169827] env[61243]: DEBUG oslo_vmware.api [None req-9617d248-0c39-4b36-b53c-acc48003a150 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339178, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.669939] env[61243]: DEBUG oslo_vmware.api [None req-9617d248-0c39-4b36-b53c-acc48003a150 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339178, 'name': PowerOffVM_Task, 'duration_secs': 0.178004} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.670243] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9617d248-0c39-4b36-b53c-acc48003a150 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 943.670479] env[61243]: DEBUG nova.compute.manager [None req-9617d248-0c39-4b36-b53c-acc48003a150 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 943.671336] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-603ff0fa-9b71-48d2-921c-a7d345f232ba {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.126661] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "7ca74bdc-c272-4b6e-a4dd-56d97144b197" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.126964] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "7ca74bdc-c272-4b6e-a4dd-56d97144b197" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.183459] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9617d248-0c39-4b36-b53c-acc48003a150 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "f77c2610-fa49-487c-9fb4-3b112f800daf" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.548s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.264469] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-783c7f42-7e77-410d-9fe2-d1b9569ee363 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.273706] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e1dba7-9296-43e0-9ad6-7d76156e7182 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.303971] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d13e2e3-5c93-436f-a5d7-e184b0d14f76 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.312193] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fdae9a4-bdee-4c58-a647-01f8849b99c5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.326252] env[61243]: DEBUG nova.compute.provider_tree [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.628881] env[61243]: DEBUG nova.compute.manager [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 944.829732] env[61243]: DEBUG nova.scheduler.client.report [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 945.123108] env[61243]: INFO nova.compute.manager [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Rebuilding instance [ 945.154979] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.162494] env[61243]: DEBUG nova.compute.manager [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.163488] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a8e3bb-41ee-4184-a142-e9678439393e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.334914] env[61243]: DEBUG oslo_concurrency.lockutils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.290s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.335496] env[61243]: DEBUG nova.compute.manager [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 945.338279] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.801s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.338523] env[61243]: DEBUG nova.objects.instance [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lazy-loading 'pci_requests' on Instance uuid c1a0d242-4eae-4c03-8341-840b41341f17 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.675202] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 945.675537] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a64b6edb-4fe1-44d0-a119-3ab6acc01f84 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.683665] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 945.683665] env[61243]: value = "task-1339179" [ 945.683665] env[61243]: _type = "Task" [ 945.683665] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.692330] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339179, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.843139] env[61243]: DEBUG nova.objects.instance [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lazy-loading 'numa_topology' on Instance uuid c1a0d242-4eae-4c03-8341-840b41341f17 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.845013] env[61243]: DEBUG nova.compute.utils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 945.846386] env[61243]: DEBUG nova.compute.manager [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 945.846555] env[61243]: DEBUG nova.network.neutron [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 945.894281] env[61243]: DEBUG nova.policy [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '785fd146024849efa0b67b1a7f731393', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9db958446334496bab832b95f2363f2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 946.142496] env[61243]: DEBUG oslo_vmware.rw_handles [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520d2904-1f16-6c39-c642-66457249f797/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 946.143941] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27480a5-a5b6-4c5b-94d6-4a0a7faa14d2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.151027] env[61243]: DEBUG oslo_vmware.rw_handles [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520d2904-1f16-6c39-c642-66457249f797/disk-0.vmdk is in state: ready. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 946.151219] env[61243]: ERROR oslo_vmware.rw_handles [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520d2904-1f16-6c39-c642-66457249f797/disk-0.vmdk due to incomplete transfer. [ 946.151467] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5d1c317e-8299-49ad-a615-d3d71f994293 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.160734] env[61243]: DEBUG oslo_vmware.rw_handles [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520d2904-1f16-6c39-c642-66457249f797/disk-0.vmdk. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 946.160995] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Uploaded image aef980b5-b4fe-49b2-97ec-338458c642cc to the Glance image server {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 946.164061] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Destroying the VM {{(pid=61243) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 946.164346] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-eb29e852-c247-4f25-9971-e0661d6943a0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.167157] env[61243]: DEBUG nova.network.neutron [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Successfully created port: 6cc2bce1-3536-4637-8cdd-4047a0131f66 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 946.171447] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 946.171447] env[61243]: value = "task-1339180" [ 946.171447] env[61243]: _type = "Task" [ 946.171447] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.181104] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339180, 'name': Destroy_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.194677] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] VM already powered off {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 946.194860] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 946.195636] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b7cd11b-d815-4448-8d51-ea9d445bad60 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.203138] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 946.203274] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-450380a6-061c-4c88-bd60-e1477f65a05d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.275358] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 946.275624] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 946.275804] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleting the datastore file [datastore2] f77c2610-fa49-487c-9fb4-3b112f800daf {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 946.276140] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28c9e011-1499-4ead-a933-3e6f7977af75 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.283372] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 946.283372] env[61243]: value = "task-1339182" [ 946.283372] env[61243]: _type = "Task" [ 946.283372] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.293385] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339182, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.347367] env[61243]: INFO nova.compute.claims [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 946.350693] env[61243]: DEBUG nova.compute.manager [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 946.681418] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339180, 'name': Destroy_Task, 'duration_secs': 0.30121} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.681739] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Destroyed the VM [ 946.681992] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Deleting Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 946.682273] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0e3d8b4d-34a8-462a-a29c-e57e6ebb52c4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.689679] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 946.689679] env[61243]: value = "task-1339183" [ 946.689679] env[61243]: _type = "Task" [ 946.689679] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.698296] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339183, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.794432] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339182, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148783} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.794718] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.794916] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 946.795118] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 947.200542] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339183, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.359698] env[61243]: DEBUG nova.compute.manager [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 947.387957] env[61243]: DEBUG nova.virt.hardware [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 947.387957] env[61243]: DEBUG nova.virt.hardware [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 947.388176] env[61243]: DEBUG nova.virt.hardware [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 947.388276] env[61243]: DEBUG nova.virt.hardware [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 947.388438] env[61243]: DEBUG nova.virt.hardware [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 947.388603] env[61243]: DEBUG nova.virt.hardware [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 947.388816] env[61243]: DEBUG nova.virt.hardware [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 947.388968] env[61243]: DEBUG nova.virt.hardware [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 947.389189] env[61243]: DEBUG nova.virt.hardware [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 947.389415] env[61243]: DEBUG nova.virt.hardware [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 947.389818] env[61243]: DEBUG nova.virt.hardware [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 947.390707] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c9ee62-a4ed-41f6-a940-5fb864c14ad7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.402193] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b775cc9c-3f2d-4785-b3a6-30c8fc7a13e4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.551199] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21bb366d-89eb-4955-9ce1-0110d814c008 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.559040] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3be021c-d594-429a-b54c-d4876dbad482 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.592505] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da46c57-19e1-4dfd-89ce-d0c56cf50c68 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.596318] env[61243]: DEBUG nova.compute.manager [req-73126a4f-6311-4a87-9bfd-24d662a3db13 req-f82a520d-e486-4c82-852f-d0d643e10ea5 service nova] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Received event network-vif-plugged-6cc2bce1-3536-4637-8cdd-4047a0131f66 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.596898] env[61243]: DEBUG oslo_concurrency.lockutils [req-73126a4f-6311-4a87-9bfd-24d662a3db13 req-f82a520d-e486-4c82-852f-d0d643e10ea5 service nova] Acquiring lock "7ce043e6-abdf-4708-80aa-221870375c23-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.597103] env[61243]: DEBUG oslo_concurrency.lockutils [req-73126a4f-6311-4a87-9bfd-24d662a3db13 req-f82a520d-e486-4c82-852f-d0d643e10ea5 service nova] Lock "7ce043e6-abdf-4708-80aa-221870375c23-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.597205] env[61243]: DEBUG oslo_concurrency.lockutils [req-73126a4f-6311-4a87-9bfd-24d662a3db13 req-f82a520d-e486-4c82-852f-d0d643e10ea5 service nova] Lock "7ce043e6-abdf-4708-80aa-221870375c23-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.597353] env[61243]: DEBUG nova.compute.manager [req-73126a4f-6311-4a87-9bfd-24d662a3db13 req-f82a520d-e486-4c82-852f-d0d643e10ea5 service nova] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] No waiting events found dispatching network-vif-plugged-6cc2bce1-3536-4637-8cdd-4047a0131f66 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 947.597526] env[61243]: WARNING nova.compute.manager [req-73126a4f-6311-4a87-9bfd-24d662a3db13 req-f82a520d-e486-4c82-852f-d0d643e10ea5 service nova] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Received unexpected event network-vif-plugged-6cc2bce1-3536-4637-8cdd-4047a0131f66 for instance with vm_state building and task_state spawning. [ 947.604490] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0061f2-2928-4dd2-90f2-17d36864db84 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.619629] env[61243]: DEBUG nova.compute.provider_tree [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.670674] env[61243]: DEBUG nova.network.neutron [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Successfully updated port: 6cc2bce1-3536-4637-8cdd-4047a0131f66 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 947.700843] env[61243]: DEBUG oslo_vmware.api [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339183, 'name': RemoveSnapshot_Task, 'duration_secs': 0.532679} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.701143] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Deleted Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 947.701412] env[61243]: INFO nova.compute.manager [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Took 13.12 seconds to snapshot the instance on the hypervisor. [ 947.824879] env[61243]: DEBUG nova.virt.hardware [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 947.825184] env[61243]: DEBUG nova.virt.hardware [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 947.825355] env[61243]: DEBUG nova.virt.hardware [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 947.825553] env[61243]: DEBUG nova.virt.hardware [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 947.825722] env[61243]: DEBUG nova.virt.hardware [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 947.825953] env[61243]: DEBUG nova.virt.hardware [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 947.826240] env[61243]: DEBUG nova.virt.hardware [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 947.826434] env[61243]: DEBUG nova.virt.hardware [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 947.826618] env[61243]: DEBUG nova.virt.hardware [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 947.826795] env[61243]: DEBUG nova.virt.hardware [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 947.826977] env[61243]: DEBUG nova.virt.hardware [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 947.827878] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-051e25fc-ca00-4722-926d-01b23aa4ad74 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.836917] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8fd4519-b4b4-4292-a1fe-f58917dd0f00 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.850496] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:72:30', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ff3ecd2f-0b10-4faf-a512-fd7a20c28df1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '67462132-75b6-4121-9a0d-e65fb2a726d2', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 947.857882] env[61243]: DEBUG oslo.service.loopingcall [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 947.858129] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 947.858349] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b1a58946-0bb2-41ff-aefe-b9cca5ea80a2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.878719] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 947.878719] env[61243]: value = "task-1339184" [ 947.878719] env[61243]: _type = "Task" [ 947.878719] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.886444] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339184, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.122781] env[61243]: DEBUG nova.scheduler.client.report [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.173865] env[61243]: DEBUG oslo_concurrency.lockutils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "refresh_cache-7ce043e6-abdf-4708-80aa-221870375c23" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.174062] env[61243]: DEBUG oslo_concurrency.lockutils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "refresh_cache-7ce043e6-abdf-4708-80aa-221870375c23" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.174230] env[61243]: DEBUG nova.network.neutron [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 948.246049] env[61243]: DEBUG nova.compute.manager [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Found 3 images (rotation: 2) {{(pid=61243) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 948.246318] env[61243]: DEBUG nova.compute.manager [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Rotating out 1 backups {{(pid=61243) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 948.246476] env[61243]: DEBUG nova.compute.manager [None req-735d8f45-ba30-4ad7-a5e6-d801b78c86ff tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Deleting image 6fe66989-aad8-47fb-9027-b19aee66207e {{(pid=61243) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 948.390056] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339184, 'name': CreateVM_Task, 'duration_secs': 0.354022} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.390435] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 948.391087] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.391234] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.391570] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 948.391851] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c4cf8a3-09db-47c6-adda-1fbae3489b0a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.397835] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 948.397835] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523be353-aa4d-33f6-72b5-1f23f795fb3d" [ 948.397835] env[61243]: _type = "Task" [ 948.397835] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.408173] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523be353-aa4d-33f6-72b5-1f23f795fb3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.628591] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.290s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.631210] env[61243]: DEBUG oslo_concurrency.lockutils [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.934s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.631458] env[61243]: DEBUG nova.objects.instance [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lazy-loading 'resources' on Instance uuid e9a1c8af-7129-4366-bda1-1c3cf40eeab1 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.680436] env[61243]: INFO nova.network.neutron [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Updating port 970e8509-164d-4ddf-8f0c-795e92883f6b with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 948.719950] env[61243]: DEBUG nova.network.neutron [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 948.909436] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523be353-aa4d-33f6-72b5-1f23f795fb3d, 'name': SearchDatastore_Task, 'duration_secs': 0.010601} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.909763] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.910017] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 948.910270] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.910768] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.910768] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 948.910927] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-960b5998-e478-4450-9777-bc2d4cf5fcdf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.919920] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 948.920123] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 948.920875] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f16ceef-dbf0-4ae9-a952-e46c672322c7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.926221] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 948.926221] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52b1d7e4-f627-0c09-888b-35e75f07343a" [ 948.926221] env[61243]: _type = "Task" [ 948.926221] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.934166] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52b1d7e4-f627-0c09-888b-35e75f07343a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.935868] env[61243]: DEBUG nova.network.neutron [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Updating instance_info_cache with network_info: [{"id": "6cc2bce1-3536-4637-8cdd-4047a0131f66", "address": "fa:16:3e:48:5d:60", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cc2bce1-35", "ovs_interfaceid": "6cc2bce1-3536-4637-8cdd-4047a0131f66", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.328687] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff1a389-0c68-48ba-94c1-68e868b7b8dc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.336747] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad95dea-3376-4ebd-8616-d191fada90cc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.368061] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69124155-beca-4a40-a4ff-68a93fd2118d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.376353] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-951d605a-a208-4836-989d-28d4846a63c6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.389752] env[61243]: DEBUG nova.compute.provider_tree [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.436960] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52b1d7e4-f627-0c09-888b-35e75f07343a, 'name': SearchDatastore_Task, 'duration_secs': 0.008417} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.437927] env[61243]: DEBUG oslo_concurrency.lockutils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "refresh_cache-7ce043e6-abdf-4708-80aa-221870375c23" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.438245] env[61243]: DEBUG nova.compute.manager [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Instance network_info: |[{"id": "6cc2bce1-3536-4637-8cdd-4047a0131f66", "address": "fa:16:3e:48:5d:60", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cc2bce1-35", "ovs_interfaceid": "6cc2bce1-3536-4637-8cdd-4047a0131f66", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 949.438486] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48f29c7d-debb-4d65-95cc-471d87b22fc5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.440908] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:5d:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a071ecf4-e713-4f97-9271-8c17952f6dee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6cc2bce1-3536-4637-8cdd-4047a0131f66', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 949.449171] env[61243]: DEBUG oslo.service.loopingcall [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 949.449406] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 949.449986] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-54060ff0-db5e-4daf-b021-62a1dfe7b93d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.465590] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 949.465590] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52820022-90a2-d68c-ca78-7c762abbb431" [ 949.465590] env[61243]: _type = "Task" [ 949.465590] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.471664] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 949.471664] env[61243]: value = "task-1339185" [ 949.471664] env[61243]: _type = "Task" [ 949.471664] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.475354] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52820022-90a2-d68c-ca78-7c762abbb431, 'name': SearchDatastore_Task, 'duration_secs': 0.017083} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.478215] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.478503] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] f77c2610-fa49-487c-9fb4-3b112f800daf/f77c2610-fa49-487c-9fb4-3b112f800daf.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 949.478757] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-586ae266-d3dc-4d04-a983-266cd3b674b2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.486209] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339185, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.487594] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 949.487594] env[61243]: value = "task-1339186" [ 949.487594] env[61243]: _type = "Task" [ 949.487594] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.495864] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339186, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.607256] env[61243]: DEBUG nova.compute.manager [req-352e2cbe-02d5-4ef4-abc8-40104be9b175 req-76b2ecdb-d990-4002-8593-4f27cae24fa2 service nova] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Received event network-changed-6cc2bce1-3536-4637-8cdd-4047a0131f66 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.607570] env[61243]: DEBUG nova.compute.manager [req-352e2cbe-02d5-4ef4-abc8-40104be9b175 req-76b2ecdb-d990-4002-8593-4f27cae24fa2 service nova] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Refreshing instance network info cache due to event network-changed-6cc2bce1-3536-4637-8cdd-4047a0131f66. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 949.608093] env[61243]: DEBUG oslo_concurrency.lockutils [req-352e2cbe-02d5-4ef4-abc8-40104be9b175 req-76b2ecdb-d990-4002-8593-4f27cae24fa2 service nova] Acquiring lock "refresh_cache-7ce043e6-abdf-4708-80aa-221870375c23" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.608093] env[61243]: DEBUG oslo_concurrency.lockutils [req-352e2cbe-02d5-4ef4-abc8-40104be9b175 req-76b2ecdb-d990-4002-8593-4f27cae24fa2 service nova] Acquired lock "refresh_cache-7ce043e6-abdf-4708-80aa-221870375c23" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.608421] env[61243]: DEBUG nova.network.neutron [req-352e2cbe-02d5-4ef4-abc8-40104be9b175 req-76b2ecdb-d990-4002-8593-4f27cae24fa2 service nova] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Refreshing network info cache for port 6cc2bce1-3536-4637-8cdd-4047a0131f66 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 949.892808] env[61243]: DEBUG nova.scheduler.client.report [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.987672] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339185, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.996582] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339186, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.090961] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fb06ad1c-5b7a-4185-8650-1d9bc5a7956d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.091312] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fb06ad1c-5b7a-4185-8650-1d9bc5a7956d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.091668] env[61243]: DEBUG nova.compute.manager [None req-fb06ad1c-5b7a-4185-8650-1d9bc5a7956d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 950.092572] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805020ff-5204-454c-929c-ab58ee8a6fe2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.101276] env[61243]: DEBUG nova.compute.manager [None req-fb06ad1c-5b7a-4185-8650-1d9bc5a7956d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61243) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 950.101863] env[61243]: DEBUG nova.objects.instance [None req-fb06ad1c-5b7a-4185-8650-1d9bc5a7956d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lazy-loading 'flavor' on Instance uuid 316572e0-c007-42cb-aaf0-3a8cfcaf24aa {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.401898] env[61243]: DEBUG oslo_concurrency.lockutils [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.771s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.404239] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.794s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.404759] env[61243]: DEBUG nova.objects.instance [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lazy-loading 'resources' on Instance uuid 09587d68-f7fd-4648-bd5c-e53b1b6464e8 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.422198] env[61243]: INFO nova.scheduler.client.report [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleted allocations for instance e9a1c8af-7129-4366-bda1-1c3cf40eeab1 [ 950.432341] env[61243]: DEBUG nova.network.neutron [req-352e2cbe-02d5-4ef4-abc8-40104be9b175 req-76b2ecdb-d990-4002-8593-4f27cae24fa2 service nova] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Updated VIF entry in instance network info cache for port 6cc2bce1-3536-4637-8cdd-4047a0131f66. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 950.432690] env[61243]: DEBUG nova.network.neutron [req-352e2cbe-02d5-4ef4-abc8-40104be9b175 req-76b2ecdb-d990-4002-8593-4f27cae24fa2 service nova] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Updating instance_info_cache with network_info: [{"id": "6cc2bce1-3536-4637-8cdd-4047a0131f66", "address": "fa:16:3e:48:5d:60", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cc2bce1-35", "ovs_interfaceid": "6cc2bce1-3536-4637-8cdd-4047a0131f66", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.435998] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.436230] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.436355] env[61243]: DEBUG nova.network.neutron [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.487271] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339185, 'name': CreateVM_Task, 'duration_secs': 0.902413} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.488151] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 950.489046] env[61243]: DEBUG oslo_concurrency.lockutils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.489277] env[61243]: DEBUG oslo_concurrency.lockutils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.489599] env[61243]: DEBUG oslo_concurrency.lockutils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 950.492829] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34ee08ad-cbc9-43d2-a95a-3b780c6ac746 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.498056] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 950.498056] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521764cf-451c-16b0-11e6-a3b2fbae4a26" [ 950.498056] env[61243]: _type = "Task" [ 950.498056] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.501315] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339186, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.578668} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.504157] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] f77c2610-fa49-487c-9fb4-3b112f800daf/f77c2610-fa49-487c-9fb4-3b112f800daf.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 950.504382] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 950.504632] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1d4ee831-e207-4bed-8c35-b24aa5b57a41 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.512420] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521764cf-451c-16b0-11e6-a3b2fbae4a26, 'name': SearchDatastore_Task, 'duration_secs': 0.009889} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.513606] env[61243]: DEBUG oslo_concurrency.lockutils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.513850] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 950.514110] env[61243]: DEBUG oslo_concurrency.lockutils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.514270] env[61243]: DEBUG oslo_concurrency.lockutils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.514456] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 950.514786] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 950.514786] env[61243]: value = "task-1339187" [ 950.514786] env[61243]: _type = "Task" [ 950.514786] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.514986] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a0d90bea-b52e-493e-8525-a6054594cca9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.524967] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339187, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.526025] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 950.526234] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 950.526973] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4339b6f-2cec-42bb-8e49-02882d112280 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.532534] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 950.532534] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ed9f11-8127-fc38-47b1-6cd27d3f0c48" [ 950.532534] env[61243]: _type = "Task" [ 950.532534] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.540588] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ed9f11-8127-fc38-47b1-6cd27d3f0c48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.607279] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb06ad1c-5b7a-4185-8650-1d9bc5a7956d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 950.607559] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f0451ae9-645e-4b8f-96af-e83c23dcec03 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.615571] env[61243]: DEBUG oslo_vmware.api [None req-fb06ad1c-5b7a-4185-8650-1d9bc5a7956d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 950.615571] env[61243]: value = "task-1339188" [ 950.615571] env[61243]: _type = "Task" [ 950.615571] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.624638] env[61243]: DEBUG oslo_vmware.api [None req-fb06ad1c-5b7a-4185-8650-1d9bc5a7956d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339188, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.931693] env[61243]: DEBUG oslo_concurrency.lockutils [None req-633a5601-705c-469f-8859-bb339331defb tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "e9a1c8af-7129-4366-bda1-1c3cf40eeab1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.257s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.934989] env[61243]: DEBUG oslo_concurrency.lockutils [req-352e2cbe-02d5-4ef4-abc8-40104be9b175 req-76b2ecdb-d990-4002-8593-4f27cae24fa2 service nova] Releasing lock "refresh_cache-7ce043e6-abdf-4708-80aa-221870375c23" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.029257] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339187, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067164} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.029586] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 951.030478] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac38ecc-3b77-482b-b1f1-aa52ad48ea85 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.044520] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ed9f11-8127-fc38-47b1-6cd27d3f0c48, 'name': SearchDatastore_Task, 'duration_secs': 0.007737} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.066035] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] f77c2610-fa49-487c-9fb4-3b112f800daf/f77c2610-fa49-487c-9fb4-3b112f800daf.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 951.070982] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-839d1718-c85f-4bd2-9e00-926cfa342588 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.073575] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-391eb1b5-26c1-458c-b614-09ac0024e23c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.093039] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 951.093039] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528aebcd-0887-f8a0-5e23-6c4bd60eb652" [ 951.093039] env[61243]: _type = "Task" [ 951.093039] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.097647] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 951.097647] env[61243]: value = "task-1339189" [ 951.097647] env[61243]: _type = "Task" [ 951.097647] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.104679] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528aebcd-0887-f8a0-5e23-6c4bd60eb652, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.112167] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339189, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.127123] env[61243]: DEBUG oslo_vmware.api [None req-fb06ad1c-5b7a-4185-8650-1d9bc5a7956d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339188, 'name': PowerOffVM_Task, 'duration_secs': 0.459569} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.130140] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb06ad1c-5b7a-4185-8650-1d9bc5a7956d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 951.130411] env[61243]: DEBUG nova.compute.manager [None req-fb06ad1c-5b7a-4185-8650-1d9bc5a7956d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 951.131594] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dff108a-be50-4d7f-96fe-d454cfa8d6e0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.171818] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d35e429-d214-400a-b828-6fd7d6c1c990 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.182535] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-599e7280-693c-455c-8715-d35342239c84 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.215575] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3abecc97-5264-477d-bb61-698f3f13f8b2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.224010] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f558e960-bf9a-47eb-bb70-23d3b51aabb6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.238268] env[61243]: DEBUG nova.compute.provider_tree [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.250589] env[61243]: DEBUG nova.network.neutron [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Updating instance_info_cache with network_info: [{"id": "970e8509-164d-4ddf-8f0c-795e92883f6b", "address": "fa:16:3e:00:af:7d", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap970e8509-16", "ovs_interfaceid": "970e8509-164d-4ddf-8f0c-795e92883f6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.605419] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528aebcd-0887-f8a0-5e23-6c4bd60eb652, 'name': SearchDatastore_Task, 'duration_secs': 0.010349} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.606178] env[61243]: DEBUG oslo_concurrency.lockutils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.606523] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 7ce043e6-abdf-4708-80aa-221870375c23/7ce043e6-abdf-4708-80aa-221870375c23.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 951.606837] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6e6a8d9b-c5e2-46af-a371-ca5fcb48233d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.611804] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339189, 'name': ReconfigVM_Task, 'duration_secs': 0.300218} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.612469] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Reconfigured VM instance instance-00000056 to attach disk [datastore2] f77c2610-fa49-487c-9fb4-3b112f800daf/f77c2610-fa49-487c-9fb4-3b112f800daf.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.613139] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ac7c1a78-7d2e-4de4-a371-4303dac1cefd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.618526] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 951.618526] env[61243]: value = "task-1339190" [ 951.618526] env[61243]: _type = "Task" [ 951.618526] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.622617] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 951.622617] env[61243]: value = "task-1339191" [ 951.622617] env[61243]: _type = "Task" [ 951.622617] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.629192] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339190, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.634854] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339191, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.637069] env[61243]: DEBUG nova.compute.manager [req-e8e91756-cb8c-40d1-95d9-a6d4d8c3e5b7 req-b2a9a02a-c591-4a1b-99c6-e91bf512d8bc service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Received event network-vif-plugged-970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.637194] env[61243]: DEBUG oslo_concurrency.lockutils [req-e8e91756-cb8c-40d1-95d9-a6d4d8c3e5b7 req-b2a9a02a-c591-4a1b-99c6-e91bf512d8bc service nova] Acquiring lock "c1a0d242-4eae-4c03-8341-840b41341f17-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.637410] env[61243]: DEBUG oslo_concurrency.lockutils [req-e8e91756-cb8c-40d1-95d9-a6d4d8c3e5b7 req-b2a9a02a-c591-4a1b-99c6-e91bf512d8bc service nova] Lock "c1a0d242-4eae-4c03-8341-840b41341f17-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.637584] env[61243]: DEBUG oslo_concurrency.lockutils [req-e8e91756-cb8c-40d1-95d9-a6d4d8c3e5b7 req-b2a9a02a-c591-4a1b-99c6-e91bf512d8bc service nova] Lock "c1a0d242-4eae-4c03-8341-840b41341f17-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.637755] env[61243]: DEBUG nova.compute.manager [req-e8e91756-cb8c-40d1-95d9-a6d4d8c3e5b7 req-b2a9a02a-c591-4a1b-99c6-e91bf512d8bc service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] No waiting events found dispatching network-vif-plugged-970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 951.637926] env[61243]: WARNING nova.compute.manager [req-e8e91756-cb8c-40d1-95d9-a6d4d8c3e5b7 req-b2a9a02a-c591-4a1b-99c6-e91bf512d8bc service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Received unexpected event network-vif-plugged-970e8509-164d-4ddf-8f0c-795e92883f6b for instance with vm_state shelved_offloaded and task_state spawning. [ 951.638113] env[61243]: DEBUG nova.compute.manager [req-e8e91756-cb8c-40d1-95d9-a6d4d8c3e5b7 req-b2a9a02a-c591-4a1b-99c6-e91bf512d8bc service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Received event network-changed-970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.638278] env[61243]: DEBUG nova.compute.manager [req-e8e91756-cb8c-40d1-95d9-a6d4d8c3e5b7 req-b2a9a02a-c591-4a1b-99c6-e91bf512d8bc service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Refreshing instance network info cache due to event network-changed-970e8509-164d-4ddf-8f0c-795e92883f6b. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 951.638453] env[61243]: DEBUG oslo_concurrency.lockutils [req-e8e91756-cb8c-40d1-95d9-a6d4d8c3e5b7 req-b2a9a02a-c591-4a1b-99c6-e91bf512d8bc service nova] Acquiring lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.643533] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fb06ad1c-5b7a-4185-8650-1d9bc5a7956d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.552s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.741806] env[61243]: DEBUG nova.scheduler.client.report [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 951.753066] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Releasing lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.756068] env[61243]: DEBUG oslo_concurrency.lockutils [req-e8e91756-cb8c-40d1-95d9-a6d4d8c3e5b7 req-b2a9a02a-c591-4a1b-99c6-e91bf512d8bc service nova] Acquired lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.756408] env[61243]: DEBUG nova.network.neutron [req-e8e91756-cb8c-40d1-95d9-a6d4d8c3e5b7 req-b2a9a02a-c591-4a1b-99c6-e91bf512d8bc service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Refreshing network info cache for port 970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 951.784078] env[61243]: DEBUG nova.virt.hardware [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='2f8fd7d49fa46f0a40fd876637fa5839',container_format='bare',created_at=2024-10-22T18:10:07Z,direct_url=,disk_format='vmdk',id=9a127f45-caa9-4072-8a3a-d6944f9047b8,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1295999899-shelved',owner='2465299214be4d5f877f340489215d6d',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-10-22T18:10:21Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 951.784442] env[61243]: DEBUG nova.virt.hardware [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 951.784648] env[61243]: DEBUG nova.virt.hardware [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.784916] env[61243]: DEBUG nova.virt.hardware [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 951.785121] env[61243]: DEBUG nova.virt.hardware [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.786306] env[61243]: DEBUG nova.virt.hardware [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 951.786306] env[61243]: DEBUG nova.virt.hardware [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 951.786306] env[61243]: DEBUG nova.virt.hardware [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 951.786306] env[61243]: DEBUG nova.virt.hardware [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 951.786306] env[61243]: DEBUG nova.virt.hardware [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 951.786621] env[61243]: DEBUG nova.virt.hardware [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.788272] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15482e00-20f7-4a55-84d0-fb2927faee50 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.800453] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b5f1cc-33db-45bd-8ecf-9cca5ea5945c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.815915] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:af:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d413776-9a8c-4afd-856f-10dbb062ca95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '970e8509-164d-4ddf-8f0c-795e92883f6b', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.823467] env[61243]: DEBUG oslo.service.loopingcall [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.823789] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 951.824046] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-095113f7-72ed-4ef2-8d3b-df7f8d60618a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.844727] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.844727] env[61243]: value = "task-1339192" [ 951.844727] env[61243]: _type = "Task" [ 951.844727] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.854407] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339192, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.131238] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339190, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.136908] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339191, 'name': Rename_Task, 'duration_secs': 0.142134} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.137253] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 952.137456] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-288d2918-b26c-457d-b943-2d6571dae1e9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.144496] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 952.144496] env[61243]: value = "task-1339193" [ 952.144496] env[61243]: _type = "Task" [ 952.144496] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.152972] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339193, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.247478] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.843s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.249524] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.245s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.251114] env[61243]: INFO nova.compute.claims [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 952.287676] env[61243]: INFO nova.scheduler.client.report [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Deleted allocations for instance 09587d68-f7fd-4648-bd5c-e53b1b6464e8 [ 952.358509] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339192, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.509279] env[61243]: DEBUG nova.network.neutron [req-e8e91756-cb8c-40d1-95d9-a6d4d8c3e5b7 req-b2a9a02a-c591-4a1b-99c6-e91bf512d8bc service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Updated VIF entry in instance network info cache for port 970e8509-164d-4ddf-8f0c-795e92883f6b. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 952.509279] env[61243]: DEBUG nova.network.neutron [req-e8e91756-cb8c-40d1-95d9-a6d4d8c3e5b7 req-b2a9a02a-c591-4a1b-99c6-e91bf512d8bc service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Updating instance_info_cache with network_info: [{"id": "970e8509-164d-4ddf-8f0c-795e92883f6b", "address": "fa:16:3e:00:af:7d", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap970e8509-16", "ovs_interfaceid": "970e8509-164d-4ddf-8f0c-795e92883f6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.515895] env[61243]: DEBUG oslo_concurrency.lockutils [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "07ee984d-476e-484d-ba80-0ec2e411faa9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.516476] env[61243]: DEBUG oslo_concurrency.lockutils [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.516906] env[61243]: DEBUG oslo_concurrency.lockutils [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "07ee984d-476e-484d-ba80-0ec2e411faa9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.517338] env[61243]: DEBUG oslo_concurrency.lockutils [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.517647] env[61243]: DEBUG oslo_concurrency.lockutils [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.520111] env[61243]: INFO nova.compute.manager [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Terminating instance [ 952.522314] env[61243]: DEBUG nova.compute.manager [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 952.522550] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 952.524064] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f85d5f1-8391-4f29-aa36-274cb94daba4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.533320] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 952.533602] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2b732e15-4f22-496f-bcd1-ecc3fe943afb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.541439] env[61243]: DEBUG oslo_vmware.api [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 952.541439] env[61243]: value = "task-1339194" [ 952.541439] env[61243]: _type = "Task" [ 952.541439] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.551585] env[61243]: DEBUG oslo_vmware.api [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339194, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.631456] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339190, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546301} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.631456] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 7ce043e6-abdf-4708-80aa-221870375c23/7ce043e6-abdf-4708-80aa-221870375c23.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 952.631973] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 952.631973] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-84b44283-5712-40aa-8e70-5980d2e82b07 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.640502] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 952.640502] env[61243]: value = "task-1339195" [ 952.640502] env[61243]: _type = "Task" [ 952.640502] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.652385] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339195, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.659892] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339193, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.798200] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1fae385e-1527-4960-bbc9-40015ed85f46 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "09587d68-f7fd-4648-bd5c-e53b1b6464e8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.736s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.856855] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339192, 'name': CreateVM_Task, 'duration_secs': 0.753723} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.857046] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 952.857715] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9a127f45-caa9-4072-8a3a-d6944f9047b8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.857896] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9a127f45-caa9-4072-8a3a-d6944f9047b8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.858311] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9a127f45-caa9-4072-8a3a-d6944f9047b8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 952.858854] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5594c744-6e3d-49ae-9124-c5242b9508bc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.865391] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 952.865391] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524d8b05-8a4f-99c4-d860-38a978fa8e40" [ 952.865391] env[61243]: _type = "Task" [ 952.865391] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.875128] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524d8b05-8a4f-99c4-d860-38a978fa8e40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.011550] env[61243]: DEBUG oslo_concurrency.lockutils [req-e8e91756-cb8c-40d1-95d9-a6d4d8c3e5b7 req-b2a9a02a-c591-4a1b-99c6-e91bf512d8bc service nova] Releasing lock "refresh_cache-c1a0d242-4eae-4c03-8341-840b41341f17" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.053975] env[61243]: DEBUG oslo_vmware.api [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339194, 'name': PowerOffVM_Task, 'duration_secs': 0.240494} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.054360] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.054482] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 953.054746] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1bed4492-6819-4245-ac4e-eaaef4039ac1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.142854] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 953.143243] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 953.143516] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Deleting the datastore file [datastore2] 07ee984d-476e-484d-ba80-0ec2e411faa9 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.146902] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d1e78dcc-45e5-4c72-9e1b-2f5e05450959 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.157980] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339193, 'name': PowerOnVM_Task, 'duration_secs': 0.78016} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.161372] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 953.161615] env[61243]: DEBUG nova.compute.manager [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 953.161964] env[61243]: DEBUG oslo_vmware.api [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 953.161964] env[61243]: value = "task-1339197" [ 953.161964] env[61243]: _type = "Task" [ 953.161964] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.162196] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339195, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083733} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.162961] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8fca250-c619-4069-ac74-ec78228a0b92 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.165480] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 953.168886] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f95269-332c-45ba-a548-892b7f7c5b1f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.176047] env[61243]: DEBUG oslo_vmware.api [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339197, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.195325] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 7ce043e6-abdf-4708-80aa-221870375c23/7ce043e6-abdf-4708-80aa-221870375c23.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 953.197642] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04e8bc5d-0e62-4145-bfef-5fbd915816a7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.219529] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 953.219529] env[61243]: value = "task-1339198" [ 953.219529] env[61243]: _type = "Task" [ 953.219529] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.232958] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339198, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.326261] env[61243]: DEBUG oslo_concurrency.lockutils [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "783d861c-8930-406e-8986-ef995de4c9dd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.326528] env[61243]: DEBUG oslo_concurrency.lockutils [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "783d861c-8930-406e-8986-ef995de4c9dd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.326743] env[61243]: DEBUG oslo_concurrency.lockutils [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "783d861c-8930-406e-8986-ef995de4c9dd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.327047] env[61243]: DEBUG oslo_concurrency.lockutils [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "783d861c-8930-406e-8986-ef995de4c9dd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.327292] env[61243]: DEBUG oslo_concurrency.lockutils [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "783d861c-8930-406e-8986-ef995de4c9dd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.329471] env[61243]: INFO nova.compute.manager [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Terminating instance [ 953.334216] env[61243]: DEBUG nova.compute.manager [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 953.334420] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 953.335741] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e2891b-ba35-4e88-a979-dd1f6d6ace61 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.344560] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.347011] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-07077624-1f46-45bf-ae43-14d35db2bc2f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.355672] env[61243]: DEBUG oslo_vmware.api [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 953.355672] env[61243]: value = "task-1339199" [ 953.355672] env[61243]: _type = "Task" [ 953.355672] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.364277] env[61243]: DEBUG oslo_vmware.api [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339199, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.377204] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9a127f45-caa9-4072-8a3a-d6944f9047b8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.377482] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Processing image 9a127f45-caa9-4072-8a3a-d6944f9047b8 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 953.377758] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9a127f45-caa9-4072-8a3a-d6944f9047b8/9a127f45-caa9-4072-8a3a-d6944f9047b8.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.377948] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9a127f45-caa9-4072-8a3a-d6944f9047b8/9a127f45-caa9-4072-8a3a-d6944f9047b8.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.378111] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.378368] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-04be9c5a-a1f5-4e56-8e21-8372c58ed8d9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.388155] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.388385] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 953.391486] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb8de53e-137a-485b-a59d-66bc2779ffa8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.398251] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 953.398251] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528685a3-dc59-3aca-26f2-5767ae4ef4b1" [ 953.398251] env[61243]: _type = "Task" [ 953.398251] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.406838] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528685a3-dc59-3aca-26f2-5767ae4ef4b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.467511] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f770d55c-72c2-4d20-8ef2-dc3f448e1c93 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.478104] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e148c3f0-5104-41c7-9587-c0ec4eedccea {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.514612] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4ae83e-f593-4635-8d15-611d7bc6d10e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.525909] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49fddaef-59d8-4359-b89e-88fe7edbdca2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.545478] env[61243]: DEBUG nova.compute.provider_tree [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.676181] env[61243]: DEBUG oslo_vmware.api [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339197, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.240728} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.676468] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.676651] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 953.676837] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 953.677034] env[61243]: INFO nova.compute.manager [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Took 1.15 seconds to destroy the instance on the hypervisor. [ 953.677293] env[61243]: DEBUG oslo.service.loopingcall [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.677489] env[61243]: DEBUG nova.compute.manager [-] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 953.677586] env[61243]: DEBUG nova.network.neutron [-] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 953.715211] env[61243]: INFO nova.compute.manager [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] bringing vm to original state: 'stopped' [ 953.729589] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339198, 'name': ReconfigVM_Task, 'duration_secs': 0.323051} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.729880] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 7ce043e6-abdf-4708-80aa-221870375c23/7ce043e6-abdf-4708-80aa-221870375c23.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 953.730531] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6924fe9f-272d-46f6-ba8c-0c549305b3fd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.738094] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 953.738094] env[61243]: value = "task-1339200" [ 953.738094] env[61243]: _type = "Task" [ 953.738094] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.750538] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339200, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.806100] env[61243]: DEBUG nova.compute.manager [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Stashing vm_state: stopped {{(pid=61243) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 953.828749] env[61243]: DEBUG oslo_concurrency.lockutils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "5103569e-a281-4e89-af69-3a070ab17564" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.829042] env[61243]: DEBUG oslo_concurrency.lockutils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "5103569e-a281-4e89-af69-3a070ab17564" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.867150] env[61243]: DEBUG oslo_vmware.api [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339199, 'name': PowerOffVM_Task, 'duration_secs': 0.234717} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.868030] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.868030] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 953.868179] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ac42ab02-22fe-44ce-bc02-3a350eafe4ec {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.910156] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Preparing fetch location {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 953.910437] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Fetch image to [datastore2] OSTACK_IMG_64445fe9-b1b7-4229-83b3-4824271407b4/OSTACK_IMG_64445fe9-b1b7-4229-83b3-4824271407b4.vmdk {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 953.910672] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Downloading stream optimized image 9a127f45-caa9-4072-8a3a-d6944f9047b8 to [datastore2] OSTACK_IMG_64445fe9-b1b7-4229-83b3-4824271407b4/OSTACK_IMG_64445fe9-b1b7-4229-83b3-4824271407b4.vmdk on the data store datastore2 as vApp {{(pid=61243) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 953.910888] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Downloading image file data 9a127f45-caa9-4072-8a3a-d6944f9047b8 to the ESX as VM named 'OSTACK_IMG_64445fe9-b1b7-4229-83b3-4824271407b4' {{(pid=61243) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 953.953463] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 953.953695] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 953.954298] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Deleting the datastore file [datastore2] 783d861c-8930-406e-8986-ef995de4c9dd {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.954298] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e9c4b3ae-1f1b-437c-8104-ecf97d3548f0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.962132] env[61243]: DEBUG oslo_vmware.api [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 953.962132] env[61243]: value = "task-1339202" [ 953.962132] env[61243]: _type = "Task" [ 953.962132] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.972448] env[61243]: DEBUG oslo_vmware.api [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339202, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.988493] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 953.988493] env[61243]: value = "resgroup-9" [ 953.988493] env[61243]: _type = "ResourcePool" [ 953.988493] env[61243]: }. {{(pid=61243) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 953.988493] env[61243]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-d070c514-1ea9-4f84-878c-bfba45e61a0b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.012035] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lease: (returnval){ [ 954.012035] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52eb7e04-0c70-a67d-4106-5c3233e275c8" [ 954.012035] env[61243]: _type = "HttpNfcLease" [ 954.012035] env[61243]: } obtained for vApp import into resource pool (val){ [ 954.012035] env[61243]: value = "resgroup-9" [ 954.012035] env[61243]: _type = "ResourcePool" [ 954.012035] env[61243]: }. {{(pid=61243) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 954.012569] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the lease: (returnval){ [ 954.012569] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52eb7e04-0c70-a67d-4106-5c3233e275c8" [ 954.012569] env[61243]: _type = "HttpNfcLease" [ 954.012569] env[61243]: } to be ready. {{(pid=61243) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 954.023282] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 954.023282] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52eb7e04-0c70-a67d-4106-5c3233e275c8" [ 954.023282] env[61243]: _type = "HttpNfcLease" [ 954.023282] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 954.050638] env[61243]: DEBUG nova.scheduler.client.report [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.250030] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339200, 'name': Rename_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.252247] env[61243]: DEBUG nova.compute.manager [req-97ed7694-c01d-48de-a997-0253bb1d4b97 req-7ca04713-bd93-4059-b63b-385a6bb27e1c service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Received event network-vif-deleted-af7195a1-9a59-4b3a-b62d-7e907c94ce7b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.252508] env[61243]: INFO nova.compute.manager [req-97ed7694-c01d-48de-a997-0253bb1d4b97 req-7ca04713-bd93-4059-b63b-385a6bb27e1c service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Neutron deleted interface af7195a1-9a59-4b3a-b62d-7e907c94ce7b; detaching it from the instance and deleting it from the info cache [ 954.252618] env[61243]: DEBUG nova.network.neutron [req-97ed7694-c01d-48de-a997-0253bb1d4b97 req-7ca04713-bd93-4059-b63b-385a6bb27e1c service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.330876] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.331691] env[61243]: DEBUG nova.compute.manager [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 954.474308] env[61243]: DEBUG oslo_vmware.api [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339202, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171761} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.474661] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 954.474869] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 954.475108] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 954.475376] env[61243]: INFO nova.compute.manager [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Took 1.14 seconds to destroy the instance on the hypervisor. [ 954.475672] env[61243]: DEBUG oslo.service.loopingcall [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.475886] env[61243]: DEBUG nova.compute.manager [-] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 954.476008] env[61243]: DEBUG nova.network.neutron [-] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 954.522312] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 954.522312] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52eb7e04-0c70-a67d-4106-5c3233e275c8" [ 954.522312] env[61243]: _type = "HttpNfcLease" [ 954.522312] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 954.556842] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.307s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.557413] env[61243]: DEBUG nova.compute.manager [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 954.561411] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.820s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.562025] env[61243]: DEBUG nova.objects.instance [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lazy-loading 'pci_requests' on Instance uuid 86179fb0-99df-4b10-a815-c19168e9521e {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.724179] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "f77c2610-fa49-487c-9fb4-3b112f800daf" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.724488] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "f77c2610-fa49-487c-9fb4-3b112f800daf" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.724658] env[61243]: DEBUG nova.compute.manager [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.725851] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2118918c-bebf-4f38-a895-2922700bccfc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.728623] env[61243]: DEBUG nova.network.neutron [-] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.735965] env[61243]: DEBUG nova.compute.manager [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61243) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 954.740300] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.740300] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-942643c5-1bc3-4ec4-b807-793a64b92ff3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.751680] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339200, 'name': Rename_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.754395] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 954.754395] env[61243]: value = "task-1339204" [ 954.754395] env[61243]: _type = "Task" [ 954.754395] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.755122] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-90b2cb4f-4bf0-457f-91cf-712a8449f51c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.767405] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339204, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.771236] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ecc272-20b4-4a76-a0bd-54a31522ec81 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.803397] env[61243]: DEBUG nova.compute.manager [req-97ed7694-c01d-48de-a997-0253bb1d4b97 req-7ca04713-bd93-4059-b63b-385a6bb27e1c service nova] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Detach interface failed, port_id=af7195a1-9a59-4b3a-b62d-7e907c94ce7b, reason: Instance 07ee984d-476e-484d-ba80-0ec2e411faa9 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 954.851351] env[61243]: DEBUG oslo_concurrency.lockutils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.021186] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 955.021186] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52eb7e04-0c70-a67d-4106-5c3233e275c8" [ 955.021186] env[61243]: _type = "HttpNfcLease" [ 955.021186] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 955.065319] env[61243]: DEBUG nova.compute.utils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 955.068481] env[61243]: DEBUG nova.objects.instance [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lazy-loading 'numa_topology' on Instance uuid 86179fb0-99df-4b10-a815-c19168e9521e {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.069625] env[61243]: DEBUG nova.compute.manager [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 955.069830] env[61243]: DEBUG nova.network.neutron [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 955.112764] env[61243]: DEBUG nova.policy [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a9e6dc85a31d4ec38efe0c750c49ead1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '113713170fb94b8a8d7985b9a02faf77', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 955.232357] env[61243]: INFO nova.compute.manager [-] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Took 1.55 seconds to deallocate network for instance. [ 955.250149] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339200, 'name': Rename_Task, 'duration_secs': 1.177457} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.252611] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 955.252611] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e460ac3-f716-4dbe-a44f-c258c9205763 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.256426] env[61243]: DEBUG nova.network.neutron [-] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.265252] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 955.265252] env[61243]: value = "task-1339205" [ 955.265252] env[61243]: _type = "Task" [ 955.265252] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.273252] env[61243]: DEBUG oslo_vmware.api [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339204, 'name': PowerOffVM_Task, 'duration_secs': 0.229} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.275246] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.275246] env[61243]: DEBUG nova.compute.manager [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 955.275562] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-738d92b0-cf77-493c-ae24-1ea93ffdc8b5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.283526] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339205, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.357830] env[61243]: DEBUG nova.network.neutron [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Successfully created port: 4a3ba85a-80c0-4d5e-a012-7d11194bb806 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 955.522655] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 955.522655] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52eb7e04-0c70-a67d-4106-5c3233e275c8" [ 955.522655] env[61243]: _type = "HttpNfcLease" [ 955.522655] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 955.572629] env[61243]: DEBUG nova.compute.manager [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 955.575890] env[61243]: INFO nova.compute.claims [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 955.739849] env[61243]: DEBUG oslo_concurrency.lockutils [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.770075] env[61243]: INFO nova.compute.manager [-] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Took 1.29 seconds to deallocate network for instance. [ 955.783940] env[61243]: DEBUG oslo_vmware.api [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339205, 'name': PowerOnVM_Task, 'duration_secs': 0.473299} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.784528] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 955.784905] env[61243]: INFO nova.compute.manager [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Took 8.43 seconds to spawn the instance on the hypervisor. [ 955.785279] env[61243]: DEBUG nova.compute.manager [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 955.786429] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-944bf07e-10ec-43c6-bccc-0c8b80aeaf5f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.793669] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "f77c2610-fa49-487c-9fb4-3b112f800daf" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.069s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.023074] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 956.023074] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52eb7e04-0c70-a67d-4106-5c3233e275c8" [ 956.023074] env[61243]: _type = "HttpNfcLease" [ 956.023074] env[61243]: } is ready. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 956.023461] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 956.023461] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52eb7e04-0c70-a67d-4106-5c3233e275c8" [ 956.023461] env[61243]: _type = "HttpNfcLease" [ 956.023461] env[61243]: }. {{(pid=61243) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 956.024207] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b694830-af6e-4212-862b-9c726b7a495f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.033502] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525f1530-01bd-35f9-8652-5a6ac68934a9/disk-0.vmdk from lease info. {{(pid=61243) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 956.033502] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525f1530-01bd-35f9-8652-5a6ac68934a9/disk-0.vmdk. {{(pid=61243) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 956.114784] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3f31ca50-5b82-48b1-8ee9-28700ccd8ef5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.277726] env[61243]: DEBUG nova.compute.manager [req-fa96aeea-a9a4-4cc0-8249-466cfddae4a6 req-22f8df16-722c-451d-afb5-8cef64222cf1 service nova] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Received event network-vif-deleted-c706f966-e43c-4d65-b536-f61daa91c8ef {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.278797] env[61243]: DEBUG oslo_concurrency.lockutils [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.316810] env[61243]: INFO nova.compute.manager [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Took 21.85 seconds to build instance. [ 956.317861] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.618242] env[61243]: DEBUG nova.compute.manager [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 956.650163] env[61243]: DEBUG nova.virt.hardware [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 956.650163] env[61243]: DEBUG nova.virt.hardware [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 956.650510] env[61243]: DEBUG nova.virt.hardware [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 956.650552] env[61243]: DEBUG nova.virt.hardware [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 956.650686] env[61243]: DEBUG nova.virt.hardware [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 956.650843] env[61243]: DEBUG nova.virt.hardware [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 956.651118] env[61243]: DEBUG nova.virt.hardware [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 956.651302] env[61243]: DEBUG nova.virt.hardware [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 956.651475] env[61243]: DEBUG nova.virt.hardware [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 956.651755] env[61243]: DEBUG nova.virt.hardware [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 956.651846] env[61243]: DEBUG nova.virt.hardware [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 956.653112] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e006a115-f0a2-4ec9-91e7-1dea5412d2d2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.664217] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c69c6eb-5322-44f6-a834-8a553208d662 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.801944] env[61243]: DEBUG oslo_concurrency.lockutils [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "f77c2610-fa49-487c-9fb4-3b112f800daf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.801944] env[61243]: DEBUG oslo_concurrency.lockutils [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "f77c2610-fa49-487c-9fb4-3b112f800daf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.802261] env[61243]: DEBUG oslo_concurrency.lockutils [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "f77c2610-fa49-487c-9fb4-3b112f800daf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.802316] env[61243]: DEBUG oslo_concurrency.lockutils [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "f77c2610-fa49-487c-9fb4-3b112f800daf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.802557] env[61243]: DEBUG oslo_concurrency.lockutils [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "f77c2610-fa49-487c-9fb4-3b112f800daf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.804944] env[61243]: INFO nova.compute.manager [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Terminating instance [ 956.806472] env[61243]: DEBUG nova.compute.manager [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 956.806660] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 956.807505] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b1d83a-d0e1-44d9-8b7b-02885d8998bd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.817865] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 956.817865] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9ec6519b-c8bf-4745-9a3d-85c8f733f57f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.818884] env[61243]: DEBUG oslo_concurrency.lockutils [None req-034795f6-fb11-4213-865d-a99e041a5ba1 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "7ce043e6-abdf-4708-80aa-221870375c23" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.361s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.849449] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7468116-13ac-43a0-94f1-f491af675703 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.859559] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f727d8f2-fc94-4f21-a649-8e45053da3f9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.906138] env[61243]: DEBUG nova.network.neutron [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Successfully updated port: 4a3ba85a-80c0-4d5e-a012-7d11194bb806 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 956.909070] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-949ba2e9-7767-4f8f-a1e5-e6663f93dcd9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.912338] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.912338] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.912489] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleting the datastore file [datastore2] f77c2610-fa49-487c-9fb4-3b112f800daf {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.915683] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f780d211-78a4-4d05-a005-c1bbd4ac3445 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.927571] env[61243]: DEBUG oslo_vmware.api [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 956.927571] env[61243]: value = "task-1339207" [ 956.927571] env[61243]: _type = "Task" [ 956.927571] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.929013] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89be07a5-f781-4a1e-a8ed-7241a6fc6808 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.950150] env[61243]: DEBUG nova.compute.provider_tree [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.957651] env[61243]: DEBUG oslo_vmware.api [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339207, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.259140] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Completed reading data from the image iterator. {{(pid=61243) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 957.259492] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525f1530-01bd-35f9-8652-5a6ac68934a9/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 957.260564] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4779a2-fdb9-4114-8fe7-22f88ed95ef8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.268252] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525f1530-01bd-35f9-8652-5a6ac68934a9/disk-0.vmdk is in state: ready. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 957.268475] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525f1530-01bd-35f9-8652-5a6ac68934a9/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 957.268760] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-f27e0564-0b9c-42a2-9311-2a594d7fafd2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.408827] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "refresh_cache-db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.409037] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "refresh_cache-db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.409205] env[61243]: DEBUG nova.network.neutron [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 957.444137] env[61243]: DEBUG oslo_vmware.api [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339207, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159289} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.444137] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 957.444137] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 957.444427] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 957.444427] env[61243]: INFO nova.compute.manager [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Took 0.64 seconds to destroy the instance on the hypervisor. [ 957.444670] env[61243]: DEBUG oslo.service.loopingcall [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.444861] env[61243]: DEBUG nova.compute.manager [-] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 957.444949] env[61243]: DEBUG nova.network.neutron [-] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 957.459156] env[61243]: DEBUG nova.scheduler.client.report [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.464822] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525f1530-01bd-35f9-8652-5a6ac68934a9/disk-0.vmdk. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 957.465048] env[61243]: INFO nova.virt.vmwareapi.images [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Downloaded image file data 9a127f45-caa9-4072-8a3a-d6944f9047b8 [ 957.465850] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06f2ec92-3058-4039-a94f-628de9637b72 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.482981] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-18b6b14c-96e6-4557-9f93-1d5d56cb4598 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.514249] env[61243]: INFO nova.virt.vmwareapi.images [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] The imported VM was unregistered [ 957.516574] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Caching image {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 957.516813] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Creating directory with path [datastore2] devstack-image-cache_base/9a127f45-caa9-4072-8a3a-d6944f9047b8 {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 957.517424] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-243035d7-4b7a-4a00-a761-1354b2001240 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.528617] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Created directory with path [datastore2] devstack-image-cache_base/9a127f45-caa9-4072-8a3a-d6944f9047b8 {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 957.528725] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_64445fe9-b1b7-4229-83b3-4824271407b4/OSTACK_IMG_64445fe9-b1b7-4229-83b3-4824271407b4.vmdk to [datastore2] devstack-image-cache_base/9a127f45-caa9-4072-8a3a-d6944f9047b8/9a127f45-caa9-4072-8a3a-d6944f9047b8.vmdk. {{(pid=61243) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 957.529027] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-726450f4-e959-4825-9ce3-2eabaf237519 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.536303] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 957.536303] env[61243]: value = "task-1339209" [ 957.536303] env[61243]: _type = "Task" [ 957.536303] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.544805] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339209, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.899446] env[61243]: INFO nova.compute.manager [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Rebuilding instance [ 957.943718] env[61243]: DEBUG nova.network.neutron [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 957.948914] env[61243]: DEBUG nova.compute.manager [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 957.950295] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77a0a1e-06c4-4799-bc40-6351159d57f8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.966122] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.405s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.967481] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.813s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.969022] env[61243]: INFO nova.compute.claims [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 958.019992] env[61243]: INFO nova.network.neutron [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Updating port f2248ec9-0230-4aa8-a016-a30ddf4f45f4 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 958.047432] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339209, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.197558] env[61243]: DEBUG nova.network.neutron [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Updating instance_info_cache with network_info: [{"id": "4a3ba85a-80c0-4d5e-a012-7d11194bb806", "address": "fa:16:3e:c9:2b:fb", "network": {"id": "0788f9ac-5a15-4431-a711-9b4f6cf79fbc", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1210957657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "113713170fb94b8a8d7985b9a02faf77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c7821ea-f92f-4f06-a4cb-05e1186a9d22", "external-id": "nsx-vlan-transportzone-69", "segmentation_id": 69, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a3ba85a-80", "ovs_interfaceid": "4a3ba85a-80c0-4d5e-a012-7d11194bb806", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.211413] env[61243]: DEBUG nova.network.neutron [-] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.315028] env[61243]: DEBUG nova.compute.manager [req-9020cbbd-755d-4df9-b5a2-107dd0c7a171 req-653ca920-bca2-4640-8bfa-e835bdc9e46d service nova] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Received event network-vif-plugged-4a3ba85a-80c0-4d5e-a012-7d11194bb806 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.315330] env[61243]: DEBUG oslo_concurrency.lockutils [req-9020cbbd-755d-4df9-b5a2-107dd0c7a171 req-653ca920-bca2-4640-8bfa-e835bdc9e46d service nova] Acquiring lock "db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.315449] env[61243]: DEBUG oslo_concurrency.lockutils [req-9020cbbd-755d-4df9-b5a2-107dd0c7a171 req-653ca920-bca2-4640-8bfa-e835bdc9e46d service nova] Lock "db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.315715] env[61243]: DEBUG oslo_concurrency.lockutils [req-9020cbbd-755d-4df9-b5a2-107dd0c7a171 req-653ca920-bca2-4640-8bfa-e835bdc9e46d service nova] Lock "db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.315813] env[61243]: DEBUG nova.compute.manager [req-9020cbbd-755d-4df9-b5a2-107dd0c7a171 req-653ca920-bca2-4640-8bfa-e835bdc9e46d service nova] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] No waiting events found dispatching network-vif-plugged-4a3ba85a-80c0-4d5e-a012-7d11194bb806 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 958.315955] env[61243]: WARNING nova.compute.manager [req-9020cbbd-755d-4df9-b5a2-107dd0c7a171 req-653ca920-bca2-4640-8bfa-e835bdc9e46d service nova] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Received unexpected event network-vif-plugged-4a3ba85a-80c0-4d5e-a012-7d11194bb806 for instance with vm_state building and task_state spawning. [ 958.316908] env[61243]: DEBUG nova.compute.manager [req-9020cbbd-755d-4df9-b5a2-107dd0c7a171 req-653ca920-bca2-4640-8bfa-e835bdc9e46d service nova] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Received event network-changed-4a3ba85a-80c0-4d5e-a012-7d11194bb806 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.317136] env[61243]: DEBUG nova.compute.manager [req-9020cbbd-755d-4df9-b5a2-107dd0c7a171 req-653ca920-bca2-4640-8bfa-e835bdc9e46d service nova] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Refreshing instance network info cache due to event network-changed-4a3ba85a-80c0-4d5e-a012-7d11194bb806. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 958.317329] env[61243]: DEBUG oslo_concurrency.lockutils [req-9020cbbd-755d-4df9-b5a2-107dd0c7a171 req-653ca920-bca2-4640-8bfa-e835bdc9e46d service nova] Acquiring lock "refresh_cache-db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.469503] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 958.469911] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2359b96e-ae26-4942-b4de-90d6b8538b26 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.483478] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 958.483478] env[61243]: value = "task-1339210" [ 958.483478] env[61243]: _type = "Task" [ 958.483478] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.501269] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339210, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.550472] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339209, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.700581] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "refresh_cache-db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.701114] env[61243]: DEBUG nova.compute.manager [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Instance network_info: |[{"id": "4a3ba85a-80c0-4d5e-a012-7d11194bb806", "address": "fa:16:3e:c9:2b:fb", "network": {"id": "0788f9ac-5a15-4431-a711-9b4f6cf79fbc", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1210957657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "113713170fb94b8a8d7985b9a02faf77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c7821ea-f92f-4f06-a4cb-05e1186a9d22", "external-id": "nsx-vlan-transportzone-69", "segmentation_id": 69, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a3ba85a-80", "ovs_interfaceid": "4a3ba85a-80c0-4d5e-a012-7d11194bb806", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 958.701471] env[61243]: DEBUG oslo_concurrency.lockutils [req-9020cbbd-755d-4df9-b5a2-107dd0c7a171 req-653ca920-bca2-4640-8bfa-e835bdc9e46d service nova] Acquired lock "refresh_cache-db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.701742] env[61243]: DEBUG nova.network.neutron [req-9020cbbd-755d-4df9-b5a2-107dd0c7a171 req-653ca920-bca2-4640-8bfa-e835bdc9e46d service nova] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Refreshing network info cache for port 4a3ba85a-80c0-4d5e-a012-7d11194bb806 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 958.703235] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:2b:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5c7821ea-f92f-4f06-a4cb-05e1186a9d22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4a3ba85a-80c0-4d5e-a012-7d11194bb806', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.711070] env[61243]: DEBUG oslo.service.loopingcall [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.712209] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 958.712647] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-605d296f-fc99-4487-b7fe-af87691d9480 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.728339] env[61243]: INFO nova.compute.manager [-] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Took 1.28 seconds to deallocate network for instance. [ 958.739708] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.739708] env[61243]: value = "task-1339211" [ 958.739708] env[61243]: _type = "Task" [ 958.739708] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.752360] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339211, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.995136] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339210, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.050746] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339209, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.188072] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03b92a0-783e-40eb-b0f5-48b1400dbe66 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.197893] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e0975e-ce17-49d8-9779-24c293cb4de4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.234864] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c250e0-cd7c-4b89-a26d-71c61dcb6a26 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.238409] env[61243]: DEBUG oslo_concurrency.lockutils [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.247854] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7b475a-eaa9-4e02-9751-b2bf92f18b46 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.257203] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339211, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.268014] env[61243]: DEBUG nova.compute.provider_tree [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.497238] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339210, 'name': PowerOffVM_Task, 'duration_secs': 0.51598} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.497676] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 959.497838] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 959.498709] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa481c81-5699-4040-a0cb-11822c28a716 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.508127] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 959.508461] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b687bc32-433b-4fd8-9fd5-e7ab398bf602 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.550231] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339209, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.561904] env[61243]: DEBUG nova.network.neutron [req-9020cbbd-755d-4df9-b5a2-107dd0c7a171 req-653ca920-bca2-4640-8bfa-e835bdc9e46d service nova] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Updated VIF entry in instance network info cache for port 4a3ba85a-80c0-4d5e-a012-7d11194bb806. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 959.562469] env[61243]: DEBUG nova.network.neutron [req-9020cbbd-755d-4df9-b5a2-107dd0c7a171 req-653ca920-bca2-4640-8bfa-e835bdc9e46d service nova] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Updating instance_info_cache with network_info: [{"id": "4a3ba85a-80c0-4d5e-a012-7d11194bb806", "address": "fa:16:3e:c9:2b:fb", "network": {"id": "0788f9ac-5a15-4431-a711-9b4f6cf79fbc", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1210957657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "113713170fb94b8a8d7985b9a02faf77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c7821ea-f92f-4f06-a4cb-05e1186a9d22", "external-id": "nsx-vlan-transportzone-69", "segmentation_id": 69, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a3ba85a-80", "ovs_interfaceid": "4a3ba85a-80c0-4d5e-a012-7d11194bb806", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.609875] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 959.610147] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 959.610348] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleting the datastore file [datastore2] 7ce043e6-abdf-4708-80aa-221870375c23 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 959.610653] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-616fecbd-e044-4eca-aaeb-3c874e1d6f6d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.618412] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 959.618412] env[61243]: value = "task-1339213" [ 959.618412] env[61243]: _type = "Task" [ 959.618412] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.628238] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339213, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.719178] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.719178] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquired lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.719178] env[61243]: DEBUG nova.network.neutron [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 959.751932] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339211, 'name': CreateVM_Task, 'duration_secs': 0.798766} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.752203] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 959.752853] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.753054] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.753747] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 959.753860] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01b562fb-a4c0-4312-ad01-329bb058569d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.759837] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 959.759837] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52111c04-ebd3-28cb-59d5-db42b6ef3039" [ 959.759837] env[61243]: _type = "Task" [ 959.759837] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.771415] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52111c04-ebd3-28cb-59d5-db42b6ef3039, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.773855] env[61243]: DEBUG nova.scheduler.client.report [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.049416] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339209, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.404158} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.049794] env[61243]: INFO nova.virt.vmwareapi.ds_util [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_64445fe9-b1b7-4229-83b3-4824271407b4/OSTACK_IMG_64445fe9-b1b7-4229-83b3-4824271407b4.vmdk to [datastore2] devstack-image-cache_base/9a127f45-caa9-4072-8a3a-d6944f9047b8/9a127f45-caa9-4072-8a3a-d6944f9047b8.vmdk. [ 960.049906] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Cleaning up location [datastore2] OSTACK_IMG_64445fe9-b1b7-4229-83b3-4824271407b4 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 960.050488] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_64445fe9-b1b7-4229-83b3-4824271407b4 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 960.050488] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9731b04b-776d-4b16-afcd-b54d0bca2f29 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.057137] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 960.057137] env[61243]: value = "task-1339214" [ 960.057137] env[61243]: _type = "Task" [ 960.057137] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.065763] env[61243]: DEBUG oslo_concurrency.lockutils [req-9020cbbd-755d-4df9-b5a2-107dd0c7a171 req-653ca920-bca2-4640-8bfa-e835bdc9e46d service nova] Releasing lock "refresh_cache-db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.066082] env[61243]: DEBUG nova.compute.manager [req-9020cbbd-755d-4df9-b5a2-107dd0c7a171 req-653ca920-bca2-4640-8bfa-e835bdc9e46d service nova] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Received event network-vif-deleted-67462132-75b6-4121-9a0d-e65fb2a726d2 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.066490] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339214, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.128460] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339213, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.385811} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.128738] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 960.128961] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 960.129306] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 960.269910] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52111c04-ebd3-28cb-59d5-db42b6ef3039, 'name': SearchDatastore_Task, 'duration_secs': 0.099362} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.270249] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.270502] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 960.270782] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.270960] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.271250] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 960.271515] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e430e83-f6c6-463c-aebb-c2c4fc32ddc5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.279252] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.279764] env[61243]: DEBUG nova.compute.manager [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 960.284144] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.957s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.285194] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 960.285383] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 960.286275] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d4ecb7a-ff9c-49ce-87db-084b6662004c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.294438] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 960.294438] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d92db1-c525-fb61-4ce4-09ec5f2ca727" [ 960.294438] env[61243]: _type = "Task" [ 960.294438] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.307450] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d92db1-c525-fb61-4ce4-09ec5f2ca727, 'name': SearchDatastore_Task, 'duration_secs': 0.009897} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.311096] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-daae1020-dc6f-4613-962d-73fe195ad837 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.317724] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 960.317724] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52000e13-c478-7fc2-8d24-e23b4aacf759" [ 960.317724] env[61243]: _type = "Task" [ 960.317724] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.327286] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52000e13-c478-7fc2-8d24-e23b4aacf759, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.344908] env[61243]: DEBUG nova.compute.manager [req-9fc01c24-4c4c-4e4f-ba8d-f3177065d03e req-2490ddee-e566-4c26-b3df-5c39228232e2 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Received event network-vif-plugged-f2248ec9-0230-4aa8-a016-a30ddf4f45f4 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.345169] env[61243]: DEBUG oslo_concurrency.lockutils [req-9fc01c24-4c4c-4e4f-ba8d-f3177065d03e req-2490ddee-e566-4c26-b3df-5c39228232e2 service nova] Acquiring lock "86179fb0-99df-4b10-a815-c19168e9521e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.345416] env[61243]: DEBUG oslo_concurrency.lockutils [req-9fc01c24-4c4c-4e4f-ba8d-f3177065d03e req-2490ddee-e566-4c26-b3df-5c39228232e2 service nova] Lock "86179fb0-99df-4b10-a815-c19168e9521e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.345630] env[61243]: DEBUG oslo_concurrency.lockutils [req-9fc01c24-4c4c-4e4f-ba8d-f3177065d03e req-2490ddee-e566-4c26-b3df-5c39228232e2 service nova] Lock "86179fb0-99df-4b10-a815-c19168e9521e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.345835] env[61243]: DEBUG nova.compute.manager [req-9fc01c24-4c4c-4e4f-ba8d-f3177065d03e req-2490ddee-e566-4c26-b3df-5c39228232e2 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] No waiting events found dispatching network-vif-plugged-f2248ec9-0230-4aa8-a016-a30ddf4f45f4 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 960.346039] env[61243]: WARNING nova.compute.manager [req-9fc01c24-4c4c-4e4f-ba8d-f3177065d03e req-2490ddee-e566-4c26-b3df-5c39228232e2 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Received unexpected event network-vif-plugged-f2248ec9-0230-4aa8-a016-a30ddf4f45f4 for instance with vm_state shelved_offloaded and task_state spawning. [ 960.346242] env[61243]: DEBUG nova.compute.manager [req-9fc01c24-4c4c-4e4f-ba8d-f3177065d03e req-2490ddee-e566-4c26-b3df-5c39228232e2 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Received event network-changed-f2248ec9-0230-4aa8-a016-a30ddf4f45f4 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.346412] env[61243]: DEBUG nova.compute.manager [req-9fc01c24-4c4c-4e4f-ba8d-f3177065d03e req-2490ddee-e566-4c26-b3df-5c39228232e2 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Refreshing instance network info cache due to event network-changed-f2248ec9-0230-4aa8-a016-a30ddf4f45f4. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 960.346608] env[61243]: DEBUG oslo_concurrency.lockutils [req-9fc01c24-4c4c-4e4f-ba8d-f3177065d03e req-2490ddee-e566-4c26-b3df-5c39228232e2 service nova] Acquiring lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.435377] env[61243]: DEBUG nova.network.neutron [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Updating instance_info_cache with network_info: [{"id": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "address": "fa:16:3e:b0:f1:66", "network": {"id": "49be247d-97e7-4085-8272-4ccf41767d10", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1335943921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "399030c8863346ea97f6da669ff3868b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2248ec9-02", "ovs_interfaceid": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.567616] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339214, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.037644} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.567889] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 960.568073] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9a127f45-caa9-4072-8a3a-d6944f9047b8/9a127f45-caa9-4072-8a3a-d6944f9047b8.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.568329] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9a127f45-caa9-4072-8a3a-d6944f9047b8/9a127f45-caa9-4072-8a3a-d6944f9047b8.vmdk to [datastore2] c1a0d242-4eae-4c03-8341-840b41341f17/c1a0d242-4eae-4c03-8341-840b41341f17.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 960.568579] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-025c554d-a2d0-4b30-bfb6-dcaa307495fe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.576059] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 960.576059] env[61243]: value = "task-1339215" [ 960.576059] env[61243]: _type = "Task" [ 960.576059] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.583873] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339215, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.786132] env[61243]: DEBUG nova.compute.utils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 960.787896] env[61243]: DEBUG nova.compute.manager [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 960.788098] env[61243]: DEBUG nova.network.neutron [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 960.795056] env[61243]: INFO nova.compute.claims [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.834037] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52000e13-c478-7fc2-8d24-e23b4aacf759, 'name': SearchDatastore_Task, 'duration_secs': 0.009936} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.834234] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.834561] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09/db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 960.835033] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7012b089-2936-4261-9431-745c39df7613 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.846359] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 960.846359] env[61243]: value = "task-1339216" [ 960.846359] env[61243]: _type = "Task" [ 960.846359] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.858830] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339216, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.937892] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Releasing lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.940883] env[61243]: DEBUG oslo_concurrency.lockutils [req-9fc01c24-4c4c-4e4f-ba8d-f3177065d03e req-2490ddee-e566-4c26-b3df-5c39228232e2 service nova] Acquired lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.941232] env[61243]: DEBUG nova.network.neutron [req-9fc01c24-4c4c-4e4f-ba8d-f3177065d03e req-2490ddee-e566-4c26-b3df-5c39228232e2 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Refreshing network info cache for port f2248ec9-0230-4aa8-a016-a30ddf4f45f4 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 960.981210] env[61243]: DEBUG nova.virt.hardware [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='0192a8fdd56da7139630ec4db522be11',container_format='bare',created_at=2024-10-22T18:10:11Z,direct_url=,disk_format='vmdk',id=0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1323857045-shelved',owner='399030c8863346ea97f6da669ff3868b',properties=ImageMetaProps,protected=,size=31666176,status='active',tags=,updated_at=2024-10-22T18:10:29Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 960.981563] env[61243]: DEBUG nova.virt.hardware [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 960.981732] env[61243]: DEBUG nova.virt.hardware [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 960.981924] env[61243]: DEBUG nova.virt.hardware [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 960.989118] env[61243]: DEBUG nova.virt.hardware [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 960.989401] env[61243]: DEBUG nova.virt.hardware [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 960.989648] env[61243]: DEBUG nova.virt.hardware [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 960.989821] env[61243]: DEBUG nova.virt.hardware [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 960.990075] env[61243]: DEBUG nova.virt.hardware [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 960.990267] env[61243]: DEBUG nova.virt.hardware [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 960.990917] env[61243]: DEBUG nova.virt.hardware [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 960.991844] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ad92ef-206d-47da-8632-18c35aadc67f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.004204] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b32dcd-4190-4b75-a693-7fb4ca1566f6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.020578] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:f1:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e547d234-640c-449b-8279-0b16f75d6627', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f2248ec9-0230-4aa8-a016-a30ddf4f45f4', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 961.030232] env[61243]: DEBUG oslo.service.loopingcall [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.030926] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 961.031220] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e77e853b-9a9d-4480-a9e5-3db02a25d2f7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.049886] env[61243]: DEBUG nova.policy [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06e247d66c8d43218302a35ead439405', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e641e72e4d2048218ef403456dbb2b82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 961.052054] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "d46fb6dc-cdd9-4688-b5e7-9efde0f770c9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.052415] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d46fb6dc-cdd9-4688-b5e7-9efde0f770c9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.052697] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "d46fb6dc-cdd9-4688-b5e7-9efde0f770c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.052929] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d46fb6dc-cdd9-4688-b5e7-9efde0f770c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.053228] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d46fb6dc-cdd9-4688-b5e7-9efde0f770c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.056827] env[61243]: INFO nova.compute.manager [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Terminating instance [ 961.059551] env[61243]: DEBUG nova.compute.manager [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 961.060075] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 961.061218] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac859b13-f9da-42f1-838d-cad628bde0bd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.066734] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 961.066734] env[61243]: value = "task-1339217" [ 961.066734] env[61243]: _type = "Task" [ 961.066734] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.074155] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 961.074940] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-13ee9f60-6a9a-4094-a3ce-2c96cc066564 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.083921] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339217, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.092960] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339215, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.094670] env[61243]: DEBUG oslo_vmware.api [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 961.094670] env[61243]: value = "task-1339218" [ 961.094670] env[61243]: _type = "Task" [ 961.094670] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.104187] env[61243]: DEBUG oslo_vmware.api [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339218, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.166081] env[61243]: DEBUG nova.virt.hardware [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 961.166516] env[61243]: DEBUG nova.virt.hardware [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 961.166886] env[61243]: DEBUG nova.virt.hardware [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 961.167292] env[61243]: DEBUG nova.virt.hardware [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 961.167541] env[61243]: DEBUG nova.virt.hardware [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 961.167902] env[61243]: DEBUG nova.virt.hardware [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 961.168329] env[61243]: DEBUG nova.virt.hardware [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 961.168629] env[61243]: DEBUG nova.virt.hardware [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 961.169007] env[61243]: DEBUG nova.virt.hardware [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 961.169333] env[61243]: DEBUG nova.virt.hardware [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 961.169667] env[61243]: DEBUG nova.virt.hardware [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 961.171224] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ebf1a3c-f382-4feb-a492-0930db28ee33 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.183196] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89e3eb2-dd10-40ed-9c96-a2e49fc2ec1a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.198665] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:5d:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a071ecf4-e713-4f97-9271-8c17952f6dee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6cc2bce1-3536-4637-8cdd-4047a0131f66', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 961.206695] env[61243]: DEBUG oslo.service.loopingcall [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.207110] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 961.207329] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-494f402a-0456-4459-91e2-d96875a2ca81 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.229519] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 961.229519] env[61243]: value = "task-1339219" [ 961.229519] env[61243]: _type = "Task" [ 961.229519] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.239311] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339219, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.302514] env[61243]: DEBUG nova.compute.manager [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 961.307948] env[61243]: INFO nova.compute.resource_tracker [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating resource usage from migration 172672ef-2a6b-494c-85d6-065c551cbfca [ 961.365354] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339216, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.379637] env[61243]: DEBUG nova.network.neutron [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Successfully created port: 1dbcd762-9bfc-4f3e-86d9-d9452a88349b {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 961.558841] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ced7fdd-76e9-40a5-a1c3-d2d043f01853 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.567830] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e83b5c3c-cf04-43ab-b725-f12a6aabec93 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.609373] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339217, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.615808] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e731e875-e820-4d0b-9473-2a0fe7006043 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.621853] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339215, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.627915] env[61243]: DEBUG oslo_vmware.api [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339218, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.631550] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da6e7e18-a7ed-4bbb-8418-10cc2bc3fa11 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.647412] env[61243]: DEBUG nova.compute.provider_tree [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.741565] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339219, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.833218] env[61243]: DEBUG nova.network.neutron [req-9fc01c24-4c4c-4e4f-ba8d-f3177065d03e req-2490ddee-e566-4c26-b3df-5c39228232e2 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Updated VIF entry in instance network info cache for port f2248ec9-0230-4aa8-a016-a30ddf4f45f4. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 961.833827] env[61243]: DEBUG nova.network.neutron [req-9fc01c24-4c4c-4e4f-ba8d-f3177065d03e req-2490ddee-e566-4c26-b3df-5c39228232e2 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Updating instance_info_cache with network_info: [{"id": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "address": "fa:16:3e:b0:f1:66", "network": {"id": "49be247d-97e7-4085-8272-4ccf41767d10", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1335943921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "399030c8863346ea97f6da669ff3868b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2248ec9-02", "ovs_interfaceid": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.863543] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339216, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.090341] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339217, 'name': CreateVM_Task, 'duration_secs': 0.876038} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.094210] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 962.094593] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339215, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.095299] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.095552] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.095953] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 962.096252] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4265f0c-576c-42a9-8769-79799db75cab {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.103109] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 962.103109] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527cbff0-2805-ee99-ea8d-a20cfea726f8" [ 962.103109] env[61243]: _type = "Task" [ 962.103109] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.113250] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527cbff0-2805-ee99-ea8d-a20cfea726f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.125163] env[61243]: DEBUG oslo_vmware.api [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339218, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.151582] env[61243]: DEBUG nova.scheduler.client.report [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.242086] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339219, 'name': CreateVM_Task, 'duration_secs': 0.906694} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.242291] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 962.243032] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.243228] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.243575] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 962.243846] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ad35448-41b1-450f-bed4-b90846c6c79c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.249918] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 962.249918] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522094bf-6173-04f6-815e-5e8f11d3ed09" [ 962.249918] env[61243]: _type = "Task" [ 962.249918] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.259085] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522094bf-6173-04f6-815e-5e8f11d3ed09, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.318522] env[61243]: DEBUG nova.compute.manager [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 962.336592] env[61243]: DEBUG oslo_concurrency.lockutils [req-9fc01c24-4c4c-4e4f-ba8d-f3177065d03e req-2490ddee-e566-4c26-b3df-5c39228232e2 service nova] Releasing lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.346467] env[61243]: DEBUG nova.virt.hardware [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.346942] env[61243]: DEBUG nova.virt.hardware [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.347135] env[61243]: DEBUG nova.virt.hardware [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.347334] env[61243]: DEBUG nova.virt.hardware [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.347488] env[61243]: DEBUG nova.virt.hardware [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.347640] env[61243]: DEBUG nova.virt.hardware [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.347856] env[61243]: DEBUG nova.virt.hardware [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.348046] env[61243]: DEBUG nova.virt.hardware [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.348237] env[61243]: DEBUG nova.virt.hardware [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.348372] env[61243]: DEBUG nova.virt.hardware [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.348584] env[61243]: DEBUG nova.virt.hardware [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.349769] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2fe732e-2a97-45bc-bd27-98b874436777 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.364687] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ea331b2-9369-4266-ade6-7b2abf1582bf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.368744] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339216, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.590828] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339215, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.617313] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.617594] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Processing image 0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 962.617853] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.617987] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.618257] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.618547] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3d10f43-c63b-42b1-9bf1-e630c073ada6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.627259] env[61243]: DEBUG oslo_vmware.api [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339218, 'name': PowerOffVM_Task, 'duration_secs': 1.058656} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.627550] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 962.627732] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 962.627995] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-612e7dbd-f5c3-4b32-9834-3ca711d5c920 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.637119] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.637341] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 962.638381] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2c44093-86d5-4fe7-9797-9c2e2141eec9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.645552] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 962.645552] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a475b1-6785-2bd4-a618-34a74e36bdb3" [ 962.645552] env[61243]: _type = "Task" [ 962.645552] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.655099] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a475b1-6785-2bd4-a618-34a74e36bdb3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.658181] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.374s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.658397] env[61243]: INFO nova.compute.manager [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Migrating [ 962.658679] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.658899] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.660377] env[61243]: DEBUG oslo_concurrency.lockutils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.809s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.662123] env[61243]: INFO nova.compute.claims [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 962.694756] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.694972] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.762866] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522094bf-6173-04f6-815e-5e8f11d3ed09, 'name': SearchDatastore_Task, 'duration_secs': 0.085463} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.763274] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.763559] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 962.763791] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.763945] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.764154] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.764434] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4a166a33-d81a-4101-b99a-e852d3b9d4bc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.781921] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 962.782214] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 962.782292] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Deleting the datastore file [datastore2] d46fb6dc-cdd9-4688-b5e7-9efde0f770c9 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 962.782593] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c40984b0-7624-4775-be48-1d1c51a6a838 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.790756] env[61243]: DEBUG oslo_vmware.api [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 962.790756] env[61243]: value = "task-1339221" [ 962.790756] env[61243]: _type = "Task" [ 962.790756] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.799711] env[61243]: DEBUG oslo_vmware.api [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339221, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.801070] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.801315] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 962.802105] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4dec1725-2315-4283-b4cc-f4b2a9f51148 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.808524] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 962.808524] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521c83d7-f9aa-63d0-a7e4-8e57266ee18b" [ 962.808524] env[61243]: _type = "Task" [ 962.808524] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.817736] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521c83d7-f9aa-63d0-a7e4-8e57266ee18b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.860102] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339216, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.092654] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339215, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.110974] env[61243]: DEBUG nova.network.neutron [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Successfully updated port: 1dbcd762-9bfc-4f3e-86d9-d9452a88349b {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 963.158910] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Preparing fetch location {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 963.159143] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Fetch image to [datastore2] OSTACK_IMG_e9dce0a0-3745-4cae-8ce3-cd9110e59e07/OSTACK_IMG_e9dce0a0-3745-4cae-8ce3-cd9110e59e07.vmdk {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 963.159245] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Downloading stream optimized image 0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e to [datastore2] OSTACK_IMG_e9dce0a0-3745-4cae-8ce3-cd9110e59e07/OSTACK_IMG_e9dce0a0-3745-4cae-8ce3-cd9110e59e07.vmdk on the data store datastore2 as vApp {{(pid=61243) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 963.159480] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Downloading image file data 0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e to the ESX as VM named 'OSTACK_IMG_e9dce0a0-3745-4cae-8ce3-cd9110e59e07' {{(pid=61243) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 963.166371] env[61243]: INFO nova.compute.rpcapi [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 963.167133] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.216158] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 963.216411] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Starting heal instance info cache {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 963.255469] env[61243]: DEBUG oslo_vmware.rw_handles [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 963.255469] env[61243]: value = "resgroup-9" [ 963.255469] env[61243]: _type = "ResourcePool" [ 963.255469] env[61243]: }. {{(pid=61243) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 963.255469] env[61243]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-337b08c9-8dfc-469d-90ff-b3477a68312f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.278074] env[61243]: DEBUG oslo_vmware.rw_handles [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lease: (returnval){ [ 963.278074] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527697e1-273b-60ba-1a0e-6827263fe17e" [ 963.278074] env[61243]: _type = "HttpNfcLease" [ 963.278074] env[61243]: } obtained for vApp import into resource pool (val){ [ 963.278074] env[61243]: value = "resgroup-9" [ 963.278074] env[61243]: _type = "ResourcePool" [ 963.278074] env[61243]: }. {{(pid=61243) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 963.278074] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the lease: (returnval){ [ 963.278074] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527697e1-273b-60ba-1a0e-6827263fe17e" [ 963.278074] env[61243]: _type = "HttpNfcLease" [ 963.278074] env[61243]: } to be ready. {{(pid=61243) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 963.281388] env[61243]: DEBUG nova.compute.manager [req-d36a647e-ce9f-484a-aab1-763446372c26 req-bcf81ddc-237f-45fd-b7b0-c0edf99a3adc service nova] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Received event network-vif-plugged-1dbcd762-9bfc-4f3e-86d9-d9452a88349b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.281388] env[61243]: DEBUG oslo_concurrency.lockutils [req-d36a647e-ce9f-484a-aab1-763446372c26 req-bcf81ddc-237f-45fd-b7b0-c0edf99a3adc service nova] Acquiring lock "7ca74bdc-c272-4b6e-a4dd-56d97144b197-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.281504] env[61243]: DEBUG oslo_concurrency.lockutils [req-d36a647e-ce9f-484a-aab1-763446372c26 req-bcf81ddc-237f-45fd-b7b0-c0edf99a3adc service nova] Lock "7ca74bdc-c272-4b6e-a4dd-56d97144b197-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.281715] env[61243]: DEBUG oslo_concurrency.lockutils [req-d36a647e-ce9f-484a-aab1-763446372c26 req-bcf81ddc-237f-45fd-b7b0-c0edf99a3adc service nova] Lock "7ca74bdc-c272-4b6e-a4dd-56d97144b197-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.281841] env[61243]: DEBUG nova.compute.manager [req-d36a647e-ce9f-484a-aab1-763446372c26 req-bcf81ddc-237f-45fd-b7b0-c0edf99a3adc service nova] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] No waiting events found dispatching network-vif-plugged-1dbcd762-9bfc-4f3e-86d9-d9452a88349b {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 963.282024] env[61243]: WARNING nova.compute.manager [req-d36a647e-ce9f-484a-aab1-763446372c26 req-bcf81ddc-237f-45fd-b7b0-c0edf99a3adc service nova] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Received unexpected event network-vif-plugged-1dbcd762-9bfc-4f3e-86d9-d9452a88349b for instance with vm_state building and task_state spawning. [ 963.289942] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 963.289942] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527697e1-273b-60ba-1a0e-6827263fe17e" [ 963.289942] env[61243]: _type = "HttpNfcLease" [ 963.289942] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 963.301093] env[61243]: DEBUG oslo_vmware.api [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339221, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.323328] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521c83d7-f9aa-63d0-a7e4-8e57266ee18b, 'name': SearchDatastore_Task, 'duration_secs': 0.133827} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.324163] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffed1004-919c-4e6b-a0d9-a10a7cc40291 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.330325] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 963.330325] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a84fd7-060c-6e54-860f-44a7db6fdbe2" [ 963.330325] env[61243]: _type = "Task" [ 963.330325] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.338770] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a84fd7-060c-6e54-860f-44a7db6fdbe2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.360556] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339216, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.399537} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.361093] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09/db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 963.361382] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 963.361747] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df3ae01d-72f1-4c72-9a68-471c64d26dc3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.370725] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 963.370725] env[61243]: value = "task-1339223" [ 963.370725] env[61243]: _type = "Task" [ 963.370725] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.379904] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339223, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.591560] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339215, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.55992} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.591882] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9a127f45-caa9-4072-8a3a-d6944f9047b8/9a127f45-caa9-4072-8a3a-d6944f9047b8.vmdk to [datastore2] c1a0d242-4eae-4c03-8341-840b41341f17/c1a0d242-4eae-4c03-8341-840b41341f17.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 963.592826] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0d53c15-f595-4879-adbd-0a5978f6db4a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.615944] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] c1a0d242-4eae-4c03-8341-840b41341f17/c1a0d242-4eae-4c03-8341-840b41341f17.vmdk or device None with type streamOptimized {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.616637] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "refresh_cache-7ca74bdc-c272-4b6e-a4dd-56d97144b197" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.616796] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "refresh_cache-7ca74bdc-c272-4b6e-a4dd-56d97144b197" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.616944] env[61243]: DEBUG nova.network.neutron [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 963.617971] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eff12ea9-ac8c-4e8d-af16-a8ba34dbc4e3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.640824] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 963.640824] env[61243]: value = "task-1339224" [ 963.640824] env[61243]: _type = "Task" [ 963.640824] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.649715] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339224, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.666916] env[61243]: DEBUG nova.network.neutron [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 963.688079] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.688335] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.688711] env[61243]: DEBUG nova.network.neutron [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 963.748596] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.749118] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquired lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.749118] env[61243]: DEBUG nova.network.neutron [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Forcefully refreshing network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 963.789161] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 963.789161] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527697e1-273b-60ba-1a0e-6827263fe17e" [ 963.789161] env[61243]: _type = "HttpNfcLease" [ 963.789161] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 963.801280] env[61243]: DEBUG oslo_vmware.api [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339221, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.630399} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.801558] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 963.801748] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 963.801929] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 963.802136] env[61243]: INFO nova.compute.manager [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Took 2.74 seconds to destroy the instance on the hypervisor. [ 963.802384] env[61243]: DEBUG oslo.service.loopingcall [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 963.802578] env[61243]: DEBUG nova.compute.manager [-] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 963.802674] env[61243]: DEBUG nova.network.neutron [-] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 963.823963] env[61243]: DEBUG nova.network.neutron [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Updating instance_info_cache with network_info: [{"id": "1dbcd762-9bfc-4f3e-86d9-d9452a88349b", "address": "fa:16:3e:34:2f:c1", "network": {"id": "bce1aecb-e043-4f68-99e2-c6b03c4bb866", "bridge": "br-int", "label": "tempest-ServersTestJSON-652904947-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e641e72e4d2048218ef403456dbb2b82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1dbcd762-9b", "ovs_interfaceid": "1dbcd762-9bfc-4f3e-86d9-d9452a88349b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.841545] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a84fd7-060c-6e54-860f-44a7db6fdbe2, 'name': SearchDatastore_Task, 'duration_secs': 0.010779} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.844571] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.844838] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 7ce043e6-abdf-4708-80aa-221870375c23/7ce043e6-abdf-4708-80aa-221870375c23.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 963.845298] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cd4a10e2-cc58-4fdf-999a-0761e7030e0b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.853136] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 963.853136] env[61243]: value = "task-1339225" [ 963.853136] env[61243]: _type = "Task" [ 963.853136] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.863962] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339225, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.880728] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339223, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076562} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.883095] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 963.884070] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3515224f-9d7d-4278-9956-93fa3a4885e6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.906681] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09/db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.907857] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2d47ea5-790a-4e23-9978-dee2e11cf67f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.923020] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1db73a-b829-410d-bfb8-026ddb9d257c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.933694] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b145a9ec-1be5-4cc5-bc62-ee2b20f5564c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.936956] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 963.936956] env[61243]: value = "task-1339226" [ 963.936956] env[61243]: _type = "Task" [ 963.936956] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.965026] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd46fca9-02ba-453f-9a76-68f5e28433c8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.970857] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339226, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.976423] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b482759-8596-4827-9a64-70f31b93e830 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.991853] env[61243]: DEBUG nova.compute.provider_tree [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.152842] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339224, 'name': ReconfigVM_Task, 'duration_secs': 0.324863} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.153301] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Reconfigured VM instance instance-00000046 to attach disk [datastore2] c1a0d242-4eae-4c03-8341-840b41341f17/c1a0d242-4eae-4c03-8341-840b41341f17.vmdk or device None with type streamOptimized {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.154753] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_type': 'disk', 'guest_format': None, 'boot_index': 0, 'encryption_format': None, 'encryption_secret_uuid': None, 'encrypted': False, 'disk_bus': None, 'encryption_options': None, 'size': 0, 'device_name': '/dev/sda', 'image_id': '6142e969-c114-4502-aa93-c018fb915a86'}], 'ephemerals': [], 'block_device_mapping': [{'guest_format': None, 'device_type': None, 'boot_index': None, 'mount_device': '/dev/sdb', 'attachment_id': '4fe790a5-0885-43c4-8f80-f814824103e5', 'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285747', 'volume_id': 'bc3b5837-80e8-4d8a-8a68-5ec6c3b89880', 'name': 'volume-bc3b5837-80e8-4d8a-8a68-5ec6c3b89880', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'c1a0d242-4eae-4c03-8341-840b41341f17', 'attached_at': '', 'detached_at': '', 'volume_id': 'bc3b5837-80e8-4d8a-8a68-5ec6c3b89880', 'serial': 'bc3b5837-80e8-4d8a-8a68-5ec6c3b89880'}, 'delete_on_termination': False, 'volume_type': None}], 'swap': None} {{(pid=61243) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 964.155067] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Volume attach. Driver type: vmdk {{(pid=61243) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 964.155237] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285747', 'volume_id': 'bc3b5837-80e8-4d8a-8a68-5ec6c3b89880', 'name': 'volume-bc3b5837-80e8-4d8a-8a68-5ec6c3b89880', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'c1a0d242-4eae-4c03-8341-840b41341f17', 'attached_at': '', 'detached_at': '', 'volume_id': 'bc3b5837-80e8-4d8a-8a68-5ec6c3b89880', 'serial': 'bc3b5837-80e8-4d8a-8a68-5ec6c3b89880'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 964.156401] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dea1a9f-5e91-4503-b77a-260ba3f6ea3f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.177023] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e103c71-20fc-477f-ae41-048fe427890c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.209765] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] volume-bc3b5837-80e8-4d8a-8a68-5ec6c3b89880/volume-bc3b5837-80e8-4d8a-8a68-5ec6c3b89880.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 964.209966] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc738425-85d2-493a-a7bf-3a5981029614 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.231049] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 964.231049] env[61243]: value = "task-1339227" [ 964.231049] env[61243]: _type = "Task" [ 964.231049] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.241355] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339227, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.284979] env[61243]: DEBUG nova.network.neutron [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 964.293128] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 964.293128] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527697e1-273b-60ba-1a0e-6827263fe17e" [ 964.293128] env[61243]: _type = "HttpNfcLease" [ 964.293128] env[61243]: } is ready. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 964.293128] env[61243]: DEBUG oslo_vmware.rw_handles [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 964.293128] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527697e1-273b-60ba-1a0e-6827263fe17e" [ 964.293128] env[61243]: _type = "HttpNfcLease" [ 964.293128] env[61243]: }. {{(pid=61243) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 964.294625] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b628405-c196-42e4-b4b3-6c491608d14a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.302898] env[61243]: DEBUG oslo_vmware.rw_handles [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528ca501-22d2-86c7-14b4-2fdcbcd42c79/disk-0.vmdk from lease info. {{(pid=61243) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 964.303033] env[61243]: DEBUG oslo_vmware.rw_handles [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Creating HTTP connection to write to file with size = 31666176 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528ca501-22d2-86c7-14b4-2fdcbcd42c79/disk-0.vmdk. {{(pid=61243) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 964.359962] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "refresh_cache-7ca74bdc-c272-4b6e-a4dd-56d97144b197" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.360321] env[61243]: DEBUG nova.compute.manager [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Instance network_info: |[{"id": "1dbcd762-9bfc-4f3e-86d9-d9452a88349b", "address": "fa:16:3e:34:2f:c1", "network": {"id": "bce1aecb-e043-4f68-99e2-c6b03c4bb866", "bridge": "br-int", "label": "tempest-ServersTestJSON-652904947-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e641e72e4d2048218ef403456dbb2b82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1dbcd762-9b", "ovs_interfaceid": "1dbcd762-9bfc-4f3e-86d9-d9452a88349b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 964.360907] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:34:2f:c1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '90878b7b-ddb7-4f47-892b-d6e06f73475f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1dbcd762-9bfc-4f3e-86d9-d9452a88349b', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 964.368600] env[61243]: DEBUG oslo.service.loopingcall [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 964.375689] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 964.379378] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7350123d-930e-4b94-ac9c-0d0511d4fc6b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.398579] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-303e622f-cb90-49c7-b4db-26f9539295c6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.404622] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339225, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45717} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.405236] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 7ce043e6-abdf-4708-80aa-221870375c23/7ce043e6-abdf-4708-80aa-221870375c23.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 964.405455] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 964.406265] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c2c71018-2843-40c0-8347-6d18ecbfd5b8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.411290] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 964.411290] env[61243]: value = "task-1339228" [ 964.411290] env[61243]: _type = "Task" [ 964.411290] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.416160] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 964.416160] env[61243]: value = "task-1339229" [ 964.416160] env[61243]: _type = "Task" [ 964.416160] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.423458] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339228, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.426764] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339229, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.448666] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339226, 'name': ReconfigVM_Task, 'duration_secs': 0.496051} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.449023] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Reconfigured VM instance instance-00000058 to attach disk [datastore2] db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09/db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.449844] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c43d1702-2384-4eff-b227-b5e9b5c23696 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.459432] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 964.459432] env[61243]: value = "task-1339230" [ 964.459432] env[61243]: _type = "Task" [ 964.459432] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.471029] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339230, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.494975] env[61243]: DEBUG nova.scheduler.client.report [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.597333] env[61243]: DEBUG nova.network.neutron [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating instance_info_cache with network_info: [{"id": "e21d7ddc-a0d3-4026-a937-f392d5e20fa4", "address": "fa:16:3e:56:77:58", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape21d7ddc-a0", "ovs_interfaceid": "e21d7ddc-a0d3-4026-a937-f392d5e20fa4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.748510] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339227, 'name': ReconfigVM_Task, 'duration_secs': 0.371238} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.751181] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Reconfigured VM instance instance-00000046 to attach disk [datastore2] volume-bc3b5837-80e8-4d8a-8a68-5ec6c3b89880/volume-bc3b5837-80e8-4d8a-8a68-5ec6c3b89880.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.762908] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2ce86d4-236a-459a-aa28-c44db11d02f1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.795998] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 964.795998] env[61243]: value = "task-1339231" [ 964.795998] env[61243]: _type = "Task" [ 964.795998] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.811014] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339231, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.874717] env[61243]: DEBUG nova.network.neutron [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.904815] env[61243]: DEBUG nova.network.neutron [-] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.925518] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339228, 'name': CreateVM_Task, 'duration_secs': 0.420992} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.927426] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 964.928577] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.928760] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.929346] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 964.932392] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e22e0b51-0a3a-4f54-99d3-850c42945be9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.934816] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339229, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074312} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.936999] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 964.938138] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-722e9501-f50f-402d-aac4-b6910e301190 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.942133] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 964.942133] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52eff5bc-1b3b-60fb-39c3-f2e89556590f" [ 964.942133] env[61243]: _type = "Task" [ 964.942133] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.967925] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 7ce043e6-abdf-4708-80aa-221870375c23/7ce043e6-abdf-4708-80aa-221870375c23.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 964.971814] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77b57c8d-2a25-4a61-94ad-a4412e336401 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.000566] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52eff5bc-1b3b-60fb-39c3-f2e89556590f, 'name': SearchDatastore_Task, 'duration_secs': 0.014483} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.000566] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.000566] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 965.000566] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.000566] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.000566] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 965.000566] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7dd43c28-cebc-4f58-b2de-2079606736f0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.005330] env[61243]: DEBUG oslo_concurrency.lockutils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.005896] env[61243]: DEBUG nova.compute.manager [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 965.008525] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339230, 'name': Rename_Task, 'duration_secs': 0.188141} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.012685] env[61243]: DEBUG oslo_concurrency.lockutils [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.273s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.012923] env[61243]: DEBUG nova.objects.instance [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lazy-loading 'resources' on Instance uuid 07ee984d-476e-484d-ba80-0ec2e411faa9 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.014091] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 965.014407] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 965.014580] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 965.015439] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 965.015439] env[61243]: value = "task-1339232" [ 965.015439] env[61243]: _type = "Task" [ 965.015439] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.015877] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a088f8ac-3192-4563-9d1b-a9e6df48326b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.017643] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73794f05-313a-4882-ac63-1b3a1c824608 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.032090] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 965.032090] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e48bde-5ebf-a985-5fc2-80e47533f25d" [ 965.032090] env[61243]: _type = "Task" [ 965.032090] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.036917] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.043772] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 965.043772] env[61243]: value = "task-1339233" [ 965.043772] env[61243]: _type = "Task" [ 965.043772] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.064118] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339233, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.067832] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e48bde-5ebf-a985-5fc2-80e47533f25d, 'name': SearchDatastore_Task, 'duration_secs': 0.028936} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.071327] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19995415-d00d-4f30-b0d4-f94d3a5088a2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.079620] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 965.079620] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5271c561-f789-93e2-8998-5813ed32413b" [ 965.079620] env[61243]: _type = "Task" [ 965.079620] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.090762] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5271c561-f789-93e2-8998-5813ed32413b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.101166] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.312315] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339231, 'name': ReconfigVM_Task, 'duration_secs': 0.196117} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.315579] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285747', 'volume_id': 'bc3b5837-80e8-4d8a-8a68-5ec6c3b89880', 'name': 'volume-bc3b5837-80e8-4d8a-8a68-5ec6c3b89880', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'c1a0d242-4eae-4c03-8341-840b41341f17', 'attached_at': '', 'detached_at': '', 'volume_id': 'bc3b5837-80e8-4d8a-8a68-5ec6c3b89880', 'serial': 'bc3b5837-80e8-4d8a-8a68-5ec6c3b89880'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 965.317353] env[61243]: DEBUG nova.compute.manager [req-4c96732b-3e62-4e4a-bc35-1436bb14ab95 req-9aa67aae-4fc2-4703-97c6-b666f1f51971 service nova] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Received event network-changed-1dbcd762-9bfc-4f3e-86d9-d9452a88349b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.317615] env[61243]: DEBUG nova.compute.manager [req-4c96732b-3e62-4e4a-bc35-1436bb14ab95 req-9aa67aae-4fc2-4703-97c6-b666f1f51971 service nova] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Refreshing instance network info cache due to event network-changed-1dbcd762-9bfc-4f3e-86d9-d9452a88349b. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 965.317791] env[61243]: DEBUG oslo_concurrency.lockutils [req-4c96732b-3e62-4e4a-bc35-1436bb14ab95 req-9aa67aae-4fc2-4703-97c6-b666f1f51971 service nova] Acquiring lock "refresh_cache-7ca74bdc-c272-4b6e-a4dd-56d97144b197" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.317950] env[61243]: DEBUG oslo_concurrency.lockutils [req-4c96732b-3e62-4e4a-bc35-1436bb14ab95 req-9aa67aae-4fc2-4703-97c6-b666f1f51971 service nova] Acquired lock "refresh_cache-7ca74bdc-c272-4b6e-a4dd-56d97144b197" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.318157] env[61243]: DEBUG nova.network.neutron [req-4c96732b-3e62-4e4a-bc35-1436bb14ab95 req-9aa67aae-4fc2-4703-97c6-b666f1f51971 service nova] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Refreshing network info cache for port 1dbcd762-9bfc-4f3e-86d9-d9452a88349b {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 965.319379] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-82f84dd3-4746-4c61-8eaa-10c4cd2fe66a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.331133] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 965.331133] env[61243]: value = "task-1339234" [ 965.331133] env[61243]: _type = "Task" [ 965.331133] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.347477] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339234, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.379376] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Releasing lock "refresh_cache-07ee984d-476e-484d-ba80-0ec2e411faa9" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.379629] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Updated the network info_cache for instance {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 965.379837] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.380137] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.380616] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.380616] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.380741] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.380907] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.381094] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61243) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 965.381898] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager.update_available_resource {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.411862] env[61243]: INFO nova.compute.manager [-] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Took 1.61 seconds to deallocate network for instance. [ 965.514182] env[61243]: DEBUG nova.compute.utils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 965.515679] env[61243]: DEBUG nova.compute.manager [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 965.515915] env[61243]: DEBUG nova.network.neutron [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 965.531775] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339232, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.562560] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339233, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.564125] env[61243]: DEBUG oslo_vmware.rw_handles [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Completed reading data from the image iterator. {{(pid=61243) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 965.564326] env[61243]: DEBUG oslo_vmware.rw_handles [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528ca501-22d2-86c7-14b4-2fdcbcd42c79/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 965.567705] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b366ad-4eda-4912-8569-69e3e07eb370 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.574570] env[61243]: DEBUG nova.policy [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cc77f359ba44dc0bc8e053d5a4b63ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3248cca8a8aa4cad84b8bd5dcf7ed0c3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 965.578984] env[61243]: DEBUG oslo_vmware.rw_handles [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528ca501-22d2-86c7-14b4-2fdcbcd42c79/disk-0.vmdk is in state: ready. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 965.579182] env[61243]: DEBUG oslo_vmware.rw_handles [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528ca501-22d2-86c7-14b4-2fdcbcd42c79/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 965.579362] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-f56ae9ea-7140-4362-922b-dea7f6cfe348 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.597064] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5271c561-f789-93e2-8998-5813ed32413b, 'name': SearchDatastore_Task, 'duration_secs': 0.012798} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.598306] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.598306] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 7ca74bdc-c272-4b6e-a4dd-56d97144b197/7ca74bdc-c272-4b6e-a4dd-56d97144b197.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 965.598306] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9fdc7a3f-0e94-4c94-8cd9-f548307df088 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.610829] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 965.610829] env[61243]: value = "task-1339235" [ 965.610829] env[61243]: _type = "Task" [ 965.610829] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.624504] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339235, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.759972] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ec04ac-7a0d-4fce-88ed-29b355a98b62 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.773568] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22580d35-b80d-41ee-9898-0a991c2ff5cf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.812291] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb1e8d1-1a05-4583-9dc0-aa5881cf34f1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.819647] env[61243]: DEBUG oslo_vmware.rw_handles [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528ca501-22d2-86c7-14b4-2fdcbcd42c79/disk-0.vmdk. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 965.819647] env[61243]: INFO nova.virt.vmwareapi.images [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Downloaded image file data 0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e [ 965.822547] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ed8b93c-f3c9-436e-968e-0b2fa7b08d85 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.828674] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0adc180-5b24-44a8-8885-74fa968f9bce {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.857872] env[61243]: DEBUG nova.compute.provider_tree [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.859576] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c0a39ba7-986f-4c94-b41e-dd1b4fe4ed50 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.865102] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339234, 'name': Rename_Task, 'duration_secs': 0.197872} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.865301] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 965.865528] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2060f867-b65e-40f0-9296-c05d088a7d8c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.874785] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 965.874785] env[61243]: value = "task-1339237" [ 965.874785] env[61243]: _type = "Task" [ 965.874785] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.885110] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.885460] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339237, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.895161] env[61243]: DEBUG nova.network.neutron [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Successfully created port: ca1babd2-ee22-456c-9b5f-205155fa93e4 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 965.921383] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.921732] env[61243]: INFO nova.virt.vmwareapi.images [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] The imported VM was unregistered [ 965.924392] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Caching image {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 965.924656] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Creating directory with path [datastore2] devstack-image-cache_base/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 965.925161] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3aa5c410-4857-4d95-b4b8-be721b219681 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.959099] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Created directory with path [datastore2] devstack-image-cache_base/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 965.959334] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_e9dce0a0-3745-4cae-8ce3-cd9110e59e07/OSTACK_IMG_e9dce0a0-3745-4cae-8ce3-cd9110e59e07.vmdk to [datastore2] devstack-image-cache_base/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e.vmdk. {{(pid=61243) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 965.959647] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-c3c8220c-40fd-4de7-bb73-82ecca54e207 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.969937] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 965.969937] env[61243]: value = "task-1339238" [ 965.969937] env[61243]: _type = "Task" [ 965.969937] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.979485] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339238, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.019771] env[61243]: DEBUG nova.compute.manager [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 966.035088] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339232, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.062867] env[61243]: DEBUG oslo_vmware.api [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339233, 'name': PowerOnVM_Task, 'duration_secs': 0.587399} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.063595] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 966.064218] env[61243]: INFO nova.compute.manager [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Took 9.45 seconds to spawn the instance on the hypervisor. [ 966.066132] env[61243]: DEBUG nova.compute.manager [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 966.067232] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66091a64-37b6-4a72-932e-835b6d695afe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.130631] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339235, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.222363] env[61243]: DEBUG nova.network.neutron [req-4c96732b-3e62-4e4a-bc35-1436bb14ab95 req-9aa67aae-4fc2-4703-97c6-b666f1f51971 service nova] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Updated VIF entry in instance network info cache for port 1dbcd762-9bfc-4f3e-86d9-d9452a88349b. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 966.222775] env[61243]: DEBUG nova.network.neutron [req-4c96732b-3e62-4e4a-bc35-1436bb14ab95 req-9aa67aae-4fc2-4703-97c6-b666f1f51971 service nova] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Updating instance_info_cache with network_info: [{"id": "1dbcd762-9bfc-4f3e-86d9-d9452a88349b", "address": "fa:16:3e:34:2f:c1", "network": {"id": "bce1aecb-e043-4f68-99e2-c6b03c4bb866", "bridge": "br-int", "label": "tempest-ServersTestJSON-652904947-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e641e72e4d2048218ef403456dbb2b82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1dbcd762-9b", "ovs_interfaceid": "1dbcd762-9bfc-4f3e-86d9-d9452a88349b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.363564] env[61243]: DEBUG nova.scheduler.client.report [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.385711] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339237, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.482020] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339238, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.536267] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339232, 'name': ReconfigVM_Task, 'duration_secs': 1.33235} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.536849] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 7ce043e6-abdf-4708-80aa-221870375c23/7ce043e6-abdf-4708-80aa-221870375c23.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 966.537555] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ff8a9b90-966e-4e30-84ae-9b293c91aeaa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.545427] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 966.545427] env[61243]: value = "task-1339239" [ 966.545427] env[61243]: _type = "Task" [ 966.545427] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.556290] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339239, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.591291] env[61243]: INFO nova.compute.manager [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Took 25.64 seconds to build instance. [ 966.624089] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339235, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.579664} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.625179] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 7ca74bdc-c272-4b6e-a4dd-56d97144b197/7ca74bdc-c272-4b6e-a4dd-56d97144b197.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 966.625412] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 966.626200] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e5a4d7-cf39-4c26-9478-10b241a4fce6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.629143] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7b189f1d-e7ea-427e-8593-46ff39a7788c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.648782] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating instance '316572e0-c007-42cb-aaf0-3a8cfcaf24aa' progress to 0 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 966.653750] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 966.653750] env[61243]: value = "task-1339240" [ 966.653750] env[61243]: _type = "Task" [ 966.653750] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.663364] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339240, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.725779] env[61243]: DEBUG oslo_concurrency.lockutils [req-4c96732b-3e62-4e4a-bc35-1436bb14ab95 req-9aa67aae-4fc2-4703-97c6-b666f1f51971 service nova] Releasing lock "refresh_cache-7ca74bdc-c272-4b6e-a4dd-56d97144b197" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.726157] env[61243]: DEBUG nova.compute.manager [req-4c96732b-3e62-4e4a-bc35-1436bb14ab95 req-9aa67aae-4fc2-4703-97c6-b666f1f51971 service nova] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Received event network-vif-deleted-bd92d46a-e97e-4169-a89e-59191cce73f3 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.869146] env[61243]: DEBUG oslo_concurrency.lockutils [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.856s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.872039] env[61243]: DEBUG oslo_concurrency.lockutils [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.593s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.872039] env[61243]: DEBUG nova.objects.instance [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lazy-loading 'resources' on Instance uuid 783d861c-8930-406e-8986-ef995de4c9dd {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.887601] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339237, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.901032] env[61243]: INFO nova.scheduler.client.report [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Deleted allocations for instance 07ee984d-476e-484d-ba80-0ec2e411faa9 [ 967.058874] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339238, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.058874] env[61243]: DEBUG nova.compute.manager [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 967.059876] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339239, 'name': Rename_Task, 'duration_secs': 0.474296} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.060325] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 967.060626] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e2c6d13-3d62-4e29-a95a-9e6abe56af35 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.072532] env[61243]: DEBUG nova.virt.hardware [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 967.072913] env[61243]: DEBUG nova.virt.hardware [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 967.073129] env[61243]: DEBUG nova.virt.hardware [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 967.073327] env[61243]: DEBUG nova.virt.hardware [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 967.073502] env[61243]: DEBUG nova.virt.hardware [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 967.073661] env[61243]: DEBUG nova.virt.hardware [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 967.073978] env[61243]: DEBUG nova.virt.hardware [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 967.074192] env[61243]: DEBUG nova.virt.hardware [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 967.074382] env[61243]: DEBUG nova.virt.hardware [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 967.074600] env[61243]: DEBUG nova.virt.hardware [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 967.074826] env[61243]: DEBUG nova.virt.hardware [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 967.075837] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c69c4b-a38b-4f3f-bb0b-0a18a5d9c393 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.080883] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 967.080883] env[61243]: value = "task-1339241" [ 967.080883] env[61243]: _type = "Task" [ 967.080883] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.090711] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab6b43a-d331-4b95-90ea-74783ae460d5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.100307] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fbd9148-c0c3-4a15-ac3f-84f5571c3c07 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.156s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.101407] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339241, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.156701] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 967.157962] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-da492130-7297-48a2-9c5e-583f553c9e5c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.169745] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339240, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081965} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.171401] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 967.171827] env[61243]: DEBUG oslo_vmware.api [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 967.171827] env[61243]: value = "task-1339242" [ 967.171827] env[61243]: _type = "Task" [ 967.171827] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.172635] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe0279c-b46c-4410-8976-2ce666c6d972 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.194395] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] VM already powered off {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 967.194745] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating instance '316572e0-c007-42cb-aaf0-3a8cfcaf24aa' progress to 17 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 967.208572] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 7ca74bdc-c272-4b6e-a4dd-56d97144b197/7ca74bdc-c272-4b6e-a4dd-56d97144b197.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 967.209409] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-328fccef-59df-48c5-a5a1-ddb40dce99e5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.232627] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 967.232627] env[61243]: value = "task-1339243" [ 967.232627] env[61243]: _type = "Task" [ 967.232627] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.243448] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339243, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.388812] env[61243]: DEBUG oslo_vmware.api [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339237, 'name': PowerOnVM_Task, 'duration_secs': 1.508804} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.389427] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 967.419096] env[61243]: DEBUG oslo_concurrency.lockutils [None req-76a865a5-7090-4835-a8bc-7cbd5a4fe838 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "07ee984d-476e-484d-ba80-0ec2e411faa9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.902s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.481349] env[61243]: DEBUG nova.compute.manager [req-fe56933a-c2ab-4253-a226-f548bced758d req-ec415c77-1097-4c35-b7dd-1078f6630152 service nova] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Received event network-vif-plugged-ca1babd2-ee22-456c-9b5f-205155fa93e4 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.481349] env[61243]: DEBUG oslo_concurrency.lockutils [req-fe56933a-c2ab-4253-a226-f548bced758d req-ec415c77-1097-4c35-b7dd-1078f6630152 service nova] Acquiring lock "5103569e-a281-4e89-af69-3a070ab17564-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.481896] env[61243]: DEBUG oslo_concurrency.lockutils [req-fe56933a-c2ab-4253-a226-f548bced758d req-ec415c77-1097-4c35-b7dd-1078f6630152 service nova] Lock "5103569e-a281-4e89-af69-3a070ab17564-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.482327] env[61243]: DEBUG oslo_concurrency.lockutils [req-fe56933a-c2ab-4253-a226-f548bced758d req-ec415c77-1097-4c35-b7dd-1078f6630152 service nova] Lock "5103569e-a281-4e89-af69-3a070ab17564-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.483821] env[61243]: DEBUG nova.compute.manager [req-fe56933a-c2ab-4253-a226-f548bced758d req-ec415c77-1097-4c35-b7dd-1078f6630152 service nova] [instance: 5103569e-a281-4e89-af69-3a070ab17564] No waiting events found dispatching network-vif-plugged-ca1babd2-ee22-456c-9b5f-205155fa93e4 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 967.483821] env[61243]: WARNING nova.compute.manager [req-fe56933a-c2ab-4253-a226-f548bced758d req-ec415c77-1097-4c35-b7dd-1078f6630152 service nova] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Received unexpected event network-vif-plugged-ca1babd2-ee22-456c-9b5f-205155fa93e4 for instance with vm_state building and task_state spawning. [ 967.494676] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339238, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.506752] env[61243]: DEBUG nova.compute.manager [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.507479] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e3c5c4-7493-4cc7-8724-d858cfcc82eb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.593543] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339241, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.594574] env[61243]: DEBUG nova.network.neutron [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Successfully updated port: ca1babd2-ee22-456c-9b5f-205155fa93e4 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 967.624014] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7593589-0712-4ef3-b9bf-da5bd360d400 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.633595] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4096b1-a777-4e15-b3ae-235ee29c3395 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.669745] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-100d7ec9-464c-4bd7-8b8b-5f6b8cc81637 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.679965] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b383bdf1-a7d6-4a91-87d1-f47c70383eaf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.694946] env[61243]: DEBUG nova.compute.provider_tree [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.711249] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 967.711558] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 967.711758] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 967.711926] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 967.712107] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 967.712269] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 967.712484] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 967.712657] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 967.712829] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 967.713093] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 967.713334] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 967.718663] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ad6ad81-01da-444a-80a0-d0b32a745d59 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.740522] env[61243]: DEBUG oslo_vmware.api [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 967.740522] env[61243]: value = "task-1339244" [ 967.740522] env[61243]: _type = "Task" [ 967.740522] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.747866] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339243, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.753828] env[61243]: DEBUG oslo_vmware.api [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339244, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.824835] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc6e434-17e8-400d-833f-0014be008c1f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.833904] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d29c3751-7219-4b60-b35e-0e91b799429a tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Suspending the VM {{(pid=61243) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 967.834196] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-e15d0e42-c5c2-449d-a0d8-2173a2f255ab {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.843747] env[61243]: DEBUG oslo_vmware.api [None req-d29c3751-7219-4b60-b35e-0e91b799429a tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 967.843747] env[61243]: value = "task-1339245" [ 967.843747] env[61243]: _type = "Task" [ 967.843747] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.856990] env[61243]: DEBUG oslo_vmware.api [None req-d29c3751-7219-4b60-b35e-0e91b799429a tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339245, 'name': SuspendVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.984145] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339238, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.028906] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2434a174-b2af-4833-af04-2e15c37d35a7 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "c1a0d242-4eae-4c03-8341-840b41341f17" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 37.807s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.093124] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339241, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.100939] env[61243]: DEBUG oslo_concurrency.lockutils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "refresh_cache-5103569e-a281-4e89-af69-3a070ab17564" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.101136] env[61243]: DEBUG oslo_concurrency.lockutils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired lock "refresh_cache-5103569e-a281-4e89-af69-3a070ab17564" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.101292] env[61243]: DEBUG nova.network.neutron [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 968.199047] env[61243]: DEBUG nova.scheduler.client.report [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.248431] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339243, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.255596] env[61243]: DEBUG oslo_vmware.api [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339244, 'name': ReconfigVM_Task, 'duration_secs': 0.362357} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.256078] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating instance '316572e0-c007-42cb-aaf0-3a8cfcaf24aa' progress to 33 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 968.356026] env[61243]: DEBUG oslo_vmware.api [None req-d29c3751-7219-4b60-b35e-0e91b799429a tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339245, 'name': SuspendVM_Task} progress is 16%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.483878] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339238, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.593092] env[61243]: DEBUG oslo_vmware.api [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339241, 'name': PowerOnVM_Task, 'duration_secs': 1.48422} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.593321] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 968.593424] env[61243]: DEBUG nova.compute.manager [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 968.594270] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454cf37e-a301-4ba3-9a10-1967fea76972 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.636768] env[61243]: DEBUG nova.network.neutron [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 968.708358] env[61243]: DEBUG oslo_concurrency.lockutils [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.837s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.712323] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 12.393s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.712607] env[61243]: DEBUG nova.objects.instance [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61243) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 968.740331] env[61243]: INFO nova.scheduler.client.report [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Deleted allocations for instance 783d861c-8930-406e-8986-ef995de4c9dd [ 968.745141] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339243, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.762576] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 968.763021] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 968.763082] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 968.763287] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 968.763511] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 968.763593] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 968.763807] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 968.764009] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 968.764222] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 968.764413] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 968.764832] env[61243]: DEBUG nova.virt.hardware [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 968.770124] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Reconfiguring VM instance instance-00000045 to detach disk 2000 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 968.773134] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d35d3178-489b-4bde-888d-8beb2177e457 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.796297] env[61243]: DEBUG oslo_vmware.api [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 968.796297] env[61243]: value = "task-1339246" [ 968.796297] env[61243]: _type = "Task" [ 968.796297] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.806657] env[61243]: DEBUG oslo_vmware.api [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339246, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.823727] env[61243]: DEBUG nova.network.neutron [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Updating instance_info_cache with network_info: [{"id": "ca1babd2-ee22-456c-9b5f-205155fa93e4", "address": "fa:16:3e:b6:8a:0d", "network": {"id": "2737f3f1-a8da-4b5f-a33d-c768f14207e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1004846100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3248cca8a8aa4cad84b8bd5dcf7ed0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca1babd2-ee", "ovs_interfaceid": "ca1babd2-ee22-456c-9b5f-205155fa93e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.856600] env[61243]: DEBUG oslo_vmware.api [None req-d29c3751-7219-4b60-b35e-0e91b799429a tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339245, 'name': SuspendVM_Task} progress is 75%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.984593] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339238, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.826683} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.984925] env[61243]: INFO nova.virt.vmwareapi.ds_util [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_e9dce0a0-3745-4cae-8ce3-cd9110e59e07/OSTACK_IMG_e9dce0a0-3745-4cae-8ce3-cd9110e59e07.vmdk to [datastore2] devstack-image-cache_base/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e.vmdk. [ 968.985143] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Cleaning up location [datastore2] OSTACK_IMG_e9dce0a0-3745-4cae-8ce3-cd9110e59e07 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 968.985266] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_e9dce0a0-3745-4cae-8ce3-cd9110e59e07 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.985557] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a231d327-f55a-4f97-be41-29dfac1be898 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.994039] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 968.994039] env[61243]: value = "task-1339247" [ 968.994039] env[61243]: _type = "Task" [ 968.994039] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.003382] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339247, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.114045] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.245299] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339243, 'name': ReconfigVM_Task, 'duration_secs': 1.522708} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.245578] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 7ca74bdc-c272-4b6e-a4dd-56d97144b197/7ca74bdc-c272-4b6e-a4dd-56d97144b197.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.246264] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1c54cb51-c383-43b2-815f-b35aa6a0fe8d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.250712] env[61243]: DEBUG oslo_concurrency.lockutils [None req-973d4eb5-4712-4986-876c-7af7ece716f7 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "783d861c-8930-406e-8986-ef995de4c9dd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.924s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.255836] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 969.255836] env[61243]: value = "task-1339248" [ 969.255836] env[61243]: _type = "Task" [ 969.255836] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.265299] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339248, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.307142] env[61243]: DEBUG oslo_vmware.api [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339246, 'name': ReconfigVM_Task, 'duration_secs': 0.175391} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.307461] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Reconfigured VM instance instance-00000045 to detach disk 2000 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 969.308255] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c53f4ef-c30d-4f60-8833-75417657be19 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.330619] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 316572e0-c007-42cb-aaf0-3a8cfcaf24aa/316572e0-c007-42cb-aaf0-3a8cfcaf24aa.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 969.331146] env[61243]: DEBUG oslo_concurrency.lockutils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Releasing lock "refresh_cache-5103569e-a281-4e89-af69-3a070ab17564" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.331473] env[61243]: DEBUG nova.compute.manager [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Instance network_info: |[{"id": "ca1babd2-ee22-456c-9b5f-205155fa93e4", "address": "fa:16:3e:b6:8a:0d", "network": {"id": "2737f3f1-a8da-4b5f-a33d-c768f14207e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1004846100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3248cca8a8aa4cad84b8bd5dcf7ed0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca1babd2-ee", "ovs_interfaceid": "ca1babd2-ee22-456c-9b5f-205155fa93e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 969.331770] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a14ce3f-dc96-4d09-ba79-73e12803a5c9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.344863] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:8a:0d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '26472e27-9835-4f87-ab7f-ca24dfee4e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ca1babd2-ee22-456c-9b5f-205155fa93e4', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 969.353178] env[61243]: DEBUG oslo.service.loopingcall [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.353178] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 969.356670] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-12f9764a-ad38-4351-a3ed-93a70fc762b5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.373033] env[61243]: DEBUG oslo_vmware.api [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 969.373033] env[61243]: value = "task-1339249" [ 969.373033] env[61243]: _type = "Task" [ 969.373033] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.379948] env[61243]: DEBUG oslo_vmware.api [None req-d29c3751-7219-4b60-b35e-0e91b799429a tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339245, 'name': SuspendVM_Task, 'duration_secs': 1.171549} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.381721] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d29c3751-7219-4b60-b35e-0e91b799429a tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Suspended the VM {{(pid=61243) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 969.381939] env[61243]: DEBUG nova.compute.manager [None req-d29c3751-7219-4b60-b35e-0e91b799429a tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 969.382216] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 969.382216] env[61243]: value = "task-1339250" [ 969.382216] env[61243]: _type = "Task" [ 969.382216] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.382975] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-828544bd-7822-486e-b341-04a04c4c368e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.394520] env[61243]: DEBUG oslo_vmware.api [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339249, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.403831] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339250, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.505475] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339247, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095486} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.505475] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.506036] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.506036] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e.vmdk to [datastore2] 86179fb0-99df-4b10-a815-c19168e9521e/86179fb0-99df-4b10-a815-c19168e9521e.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 969.506333] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e777cf3-68d8-4662-8cae-c942c95073c2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.514457] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 969.514457] env[61243]: value = "task-1339251" [ 969.514457] env[61243]: _type = "Task" [ 969.514457] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.523418] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339251, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.668667] env[61243]: DEBUG nova.compute.manager [req-96819f53-c6d8-4c4c-8480-9a7b6623120f req-85e2e6f6-0041-4dc2-b564-be601cf303d5 service nova] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Received event network-changed-ca1babd2-ee22-456c-9b5f-205155fa93e4 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.668936] env[61243]: DEBUG nova.compute.manager [req-96819f53-c6d8-4c4c-8480-9a7b6623120f req-85e2e6f6-0041-4dc2-b564-be601cf303d5 service nova] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Refreshing instance network info cache due to event network-changed-ca1babd2-ee22-456c-9b5f-205155fa93e4. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 969.669214] env[61243]: DEBUG oslo_concurrency.lockutils [req-96819f53-c6d8-4c4c-8480-9a7b6623120f req-85e2e6f6-0041-4dc2-b564-be601cf303d5 service nova] Acquiring lock "refresh_cache-5103569e-a281-4e89-af69-3a070ab17564" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.669425] env[61243]: DEBUG oslo_concurrency.lockutils [req-96819f53-c6d8-4c4c-8480-9a7b6623120f req-85e2e6f6-0041-4dc2-b564-be601cf303d5 service nova] Acquired lock "refresh_cache-5103569e-a281-4e89-af69-3a070ab17564" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.669700] env[61243]: DEBUG nova.network.neutron [req-96819f53-c6d8-4c4c-8480-9a7b6623120f req-85e2e6f6-0041-4dc2-b564-be601cf303d5 service nova] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Refreshing network info cache for port ca1babd2-ee22-456c-9b5f-205155fa93e4 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 969.726200] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cace7bfe-e684-416d-8b5d-401ecc1c99e5 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.727340] env[61243]: DEBUG oslo_concurrency.lockutils [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.489s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.727543] env[61243]: DEBUG nova.objects.instance [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lazy-loading 'resources' on Instance uuid f77c2610-fa49-487c-9fb4-3b112f800daf {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.767056] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339248, 'name': Rename_Task, 'duration_secs': 0.155748} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.767355] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 969.767618] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b1fcb34-07f5-4144-8ec0-e97423bc86cf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.775737] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 969.775737] env[61243]: value = "task-1339252" [ 969.775737] env[61243]: _type = "Task" [ 969.775737] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.784584] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339252, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.884319] env[61243]: DEBUG oslo_vmware.api [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339249, 'name': ReconfigVM_Task, 'duration_secs': 0.290601} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.884622] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 316572e0-c007-42cb-aaf0-3a8cfcaf24aa/316572e0-c007-42cb-aaf0-3a8cfcaf24aa.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.884952] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating instance '316572e0-c007-42cb-aaf0-3a8cfcaf24aa' progress to 50 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 969.897063] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339250, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.026147] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339251, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.290178] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339252, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.364260] env[61243]: DEBUG oslo_concurrency.lockutils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "55f24903-b0b4-4124-afde-f31b2680a53a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.364524] env[61243]: DEBUG oslo_concurrency.lockutils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.395408] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de648fd3-b6c4-41cb-8160-dfc1bd5ca90e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.405822] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339250, 'name': CreateVM_Task, 'duration_secs': 0.803293} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.405911] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 970.406656] env[61243]: DEBUG oslo_concurrency.lockutils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.406835] env[61243]: DEBUG oslo_concurrency.lockutils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.407187] env[61243]: DEBUG oslo_concurrency.lockutils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 970.407457] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bde565ed-db94-43af-a07f-8dac599dfad7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.426491] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "7ce043e6-abdf-4708-80aa-221870375c23" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.426764] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "7ce043e6-abdf-4708-80aa-221870375c23" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.426984] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "7ce043e6-abdf-4708-80aa-221870375c23-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.427188] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "7ce043e6-abdf-4708-80aa-221870375c23-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.427363] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "7ce043e6-abdf-4708-80aa-221870375c23-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.432612] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ebb612-6ff1-425e-8cf7-8433e4000b12 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.436291] env[61243]: INFO nova.compute.manager [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Terminating instance [ 970.440145] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 970.440145] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529efef9-c31e-7ad3-1b1e-9b6803707436" [ 970.440145] env[61243]: _type = "Task" [ 970.440145] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.440778] env[61243]: DEBUG nova.compute.manager [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 970.441013] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 970.442299] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efcb9662-3e34-4816-bacc-91bce16b8529 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.464332] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.464332] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.464332] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.464332] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.464332] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.464332] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating instance '316572e0-c007-42cb-aaf0-3a8cfcaf24aa' progress to 67 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 970.467365] env[61243]: INFO nova.compute.manager [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Terminating instance [ 970.474021] env[61243]: DEBUG nova.compute.manager [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 970.474021] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 970.478018] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784e021c-f5f3-4349-9d19-3a0413eb005b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.482592] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 970.482898] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529efef9-c31e-7ad3-1b1e-9b6803707436, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.483563] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-33bec23e-f350-4dfa-b7e6-e6180fb834ec {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.486976] env[61243]: DEBUG nova.network.neutron [req-96819f53-c6d8-4c4c-8480-9a7b6623120f req-85e2e6f6-0041-4dc2-b564-be601cf303d5 service nova] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Updated VIF entry in instance network info cache for port ca1babd2-ee22-456c-9b5f-205155fa93e4. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 970.487352] env[61243]: DEBUG nova.network.neutron [req-96819f53-c6d8-4c4c-8480-9a7b6623120f req-85e2e6f6-0041-4dc2-b564-be601cf303d5 service nova] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Updating instance_info_cache with network_info: [{"id": "ca1babd2-ee22-456c-9b5f-205155fa93e4", "address": "fa:16:3e:b6:8a:0d", "network": {"id": "2737f3f1-a8da-4b5f-a33d-c768f14207e2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1004846100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3248cca8a8aa4cad84b8bd5dcf7ed0c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca1babd2-ee", "ovs_interfaceid": "ca1babd2-ee22-456c-9b5f-205155fa93e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.490755] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 970.492552] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1e6fe2c-f093-43e0-aaa2-e99d2ad702c4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.494444] env[61243]: DEBUG oslo_vmware.api [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 970.494444] env[61243]: value = "task-1339253" [ 970.494444] env[61243]: _type = "Task" [ 970.494444] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.507898] env[61243]: DEBUG oslo_vmware.api [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339253, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.516418] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a443be-2b14-4728-9d13-1acb578daf94 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.534870] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339251, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.537485] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56652868-6a80-4be8-b07a-1d31efcfad5a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.572015] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a739a0-998f-4c04-ae07-160bef154621 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.580938] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72c6af4-c471-4eba-b4b7-da7bf0e4cfd3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.596755] env[61243]: DEBUG nova.compute.provider_tree [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 970.635808] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 970.636059] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 970.636246] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Deleting the datastore file [datastore2] db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.637130] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f670a4a-ace8-4dba-96f4-08950a86a0ac {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.645895] env[61243]: DEBUG oslo_vmware.api [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 970.645895] env[61243]: value = "task-1339255" [ 970.645895] env[61243]: _type = "Task" [ 970.645895] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.656103] env[61243]: DEBUG oslo_vmware.api [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339255, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.787287] env[61243]: DEBUG oslo_vmware.api [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339252, 'name': PowerOnVM_Task, 'duration_secs': 0.836352} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.787593] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 970.787806] env[61243]: INFO nova.compute.manager [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Took 8.47 seconds to spawn the instance on the hypervisor. [ 970.787996] env[61243]: DEBUG nova.compute.manager [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.788841] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec3b7e49-2998-45b3-8c3c-e049a910df91 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.867611] env[61243]: DEBUG nova.compute.manager [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 970.953433] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529efef9-c31e-7ad3-1b1e-9b6803707436, 'name': SearchDatastore_Task, 'duration_secs': 0.099356} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.953774] env[61243]: DEBUG oslo_concurrency.lockutils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.954144] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 970.954306] env[61243]: DEBUG oslo_concurrency.lockutils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.954461] env[61243]: DEBUG oslo_concurrency.lockutils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.954662] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 970.954978] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4f42bf16-8f5f-4080-8f12-d64690645798 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.970430] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 970.970588] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 970.971425] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65ec7ba9-340e-4b5b-8509-221a27a085b6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.983446] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 970.983446] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5266955e-525d-ebbb-17c0-dc6b626e5e4b" [ 970.983446] env[61243]: _type = "Task" [ 970.983446] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.992243] env[61243]: DEBUG oslo_concurrency.lockutils [req-96819f53-c6d8-4c4c-8480-9a7b6623120f req-85e2e6f6-0041-4dc2-b564-be601cf303d5 service nova] Releasing lock "refresh_cache-5103569e-a281-4e89-af69-3a070ab17564" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.992530] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5266955e-525d-ebbb-17c0-dc6b626e5e4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.996794] env[61243]: DEBUG oslo_concurrency.lockutils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "5660de64-7152-4832-960e-2deb1c0bc37e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.997050] env[61243]: DEBUG oslo_concurrency.lockutils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "5660de64-7152-4832-960e-2deb1c0bc37e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.008692] env[61243]: DEBUG oslo_vmware.api [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339253, 'name': PowerOffVM_Task, 'duration_secs': 0.472417} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.009579] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 971.009803] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 971.010092] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-75de6226-d7b1-49b9-9004-41623b475faa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.019719] env[61243]: DEBUG nova.network.neutron [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Port e21d7ddc-a0d3-4026-a937-f392d5e20fa4 binding to destination host cpu-1 is already ACTIVE {{(pid=61243) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 971.027606] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339251, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.119983] env[61243]: ERROR nova.scheduler.client.report [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [req-02911071-b0b7-49e6-b4e7-eeb91b7dc13a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-02911071-b0b7-49e6-b4e7-eeb91b7dc13a"}]} [ 971.144606] env[61243]: DEBUG nova.scheduler.client.report [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 971.156619] env[61243]: DEBUG oslo_vmware.api [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339255, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.162601] env[61243]: DEBUG nova.scheduler.client.report [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 971.162948] env[61243]: DEBUG nova.compute.provider_tree [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 971.176319] env[61243]: DEBUG nova.scheduler.client.report [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 971.196600] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 971.197035] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 971.197481] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleting the datastore file [datastore2] 7ce043e6-abdf-4708-80aa-221870375c23 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 971.198495] env[61243]: DEBUG nova.scheduler.client.report [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 971.200894] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-572ae6b0-95af-4b90-ba61-28beea73fb40 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.209329] env[61243]: DEBUG oslo_vmware.api [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 971.209329] env[61243]: value = "task-1339257" [ 971.209329] env[61243]: _type = "Task" [ 971.209329] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.219142] env[61243]: DEBUG oslo_vmware.api [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339257, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.306525] env[61243]: INFO nova.compute.manager [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Took 26.17 seconds to build instance. [ 971.390642] env[61243]: DEBUG oslo_concurrency.lockutils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.412955] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18dfdb5d-8fd1-43a7-a7cf-090a91bc1b72 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.421542] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba472b9-f53a-45b8-8f3b-48a756ba24d4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.455120] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f98b7c-60db-46d6-9827-80c987750ad5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.464396] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f461f536-1305-47d5-b076-b125e1875fb3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.481750] env[61243]: DEBUG nova.compute.provider_tree [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 971.498639] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5266955e-525d-ebbb-17c0-dc6b626e5e4b, 'name': SearchDatastore_Task, 'duration_secs': 0.09536} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.499799] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75a4f7b1-7a0b-4797-b920-c59ce450f314 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.502515] env[61243]: DEBUG nova.compute.manager [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 971.510448] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 971.510448] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52763f76-2d68-b8f1-8993-d49f1b38b7a2" [ 971.510448] env[61243]: _type = "Task" [ 971.510448] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.520210] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52763f76-2d68-b8f1-8993-d49f1b38b7a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.534062] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339251, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.659045] env[61243]: DEBUG oslo_vmware.api [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339255, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.719304] env[61243]: DEBUG oslo_vmware.api [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339257, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.811201] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e028412e-5d2d-4bab-8e0f-4760c7fdc794 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "7ca74bdc-c272-4b6e-a4dd-56d97144b197" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.684s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.019317] env[61243]: DEBUG nova.scheduler.client.report [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 121 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 972.019618] env[61243]: DEBUG nova.compute.provider_tree [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 121 to 122 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 972.019811] env[61243]: DEBUG nova.compute.provider_tree [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 972.024403] env[61243]: DEBUG oslo_concurrency.lockutils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.038997] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52763f76-2d68-b8f1-8993-d49f1b38b7a2, 'name': SearchDatastore_Task, 'duration_secs': 0.092687} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.043440] env[61243]: DEBUG oslo_concurrency.lockutils [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.316s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.045645] env[61243]: DEBUG oslo_concurrency.lockutils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.045940] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 5103569e-a281-4e89-af69-3a070ab17564/5103569e-a281-4e89-af69-3a070ab17564.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 972.049867] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.165s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.050071] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.050236] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61243) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 972.050680] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.129s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.050776] env[61243]: DEBUG nova.objects.instance [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lazy-loading 'resources' on Instance uuid d46fb6dc-cdd9-4688-b5e7-9efde0f770c9 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.052971] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.053254] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.053511] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.054791] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3d677084-6ee8-450c-b890-79c7294ced91 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.058844] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd592f60-8007-44b6-83e1-78a35afa9fc9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.067956] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339251, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.073337] env[61243]: INFO nova.scheduler.client.report [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleted allocations for instance f77c2610-fa49-487c-9fb4-3b112f800daf [ 972.081421] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 972.081421] env[61243]: value = "task-1339258" [ 972.081421] env[61243]: _type = "Task" [ 972.081421] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.083163] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e18c8e39-9a74-4013-9341-7b0ddaeb5691 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.101190] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339258, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.112101] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87435165-da5a-406e-b8ba-e9aa6c19b8af {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.122109] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af0e106-2867-4e9b-9dff-ec5bfed645f2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.153973] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180071MB free_disk=153GB free_vcpus=48 pci_devices=None {{(pid=61243) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 972.154202] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.164579] env[61243]: DEBUG oslo_vmware.api [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339255, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.220710] env[61243]: DEBUG oslo_vmware.api [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339257, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.533481] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339251, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.589863] env[61243]: DEBUG oslo_concurrency.lockutils [None req-88f8db5d-0e2d-467f-aec7-d7c833c62d29 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "f77c2610-fa49-487c-9fb4-3b112f800daf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.788s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.596644] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339258, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.666798] env[61243]: DEBUG oslo_vmware.api [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339255, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.89141} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.669263] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.669464] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 972.669675] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 972.669857] env[61243]: INFO nova.compute.manager [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Took 2.20 seconds to destroy the instance on the hypervisor. [ 972.670116] env[61243]: DEBUG oslo.service.loopingcall [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.670525] env[61243]: DEBUG nova.compute.manager [-] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 972.670625] env[61243]: DEBUG nova.network.neutron [-] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 972.722970] env[61243]: DEBUG oslo_vmware.api [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339257, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.750540] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed06179-5d54-44ef-9164-224413bbb4f6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.759282] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e99f6fa4-1d8b-42c5-916a-6dc0216178bf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.789779] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-263f30dd-258c-4e25-a78d-54aa8517dfed {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.799892] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de7a3629-84e2-448a-8ad0-7eb4761c0167 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.816622] env[61243]: DEBUG nova.compute.provider_tree [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 972.876963] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "818476f0-e928-44d5-b3d4-101b7a05c9aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.877233] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "818476f0-e928-44d5-b3d4-101b7a05c9aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.966611] env[61243]: DEBUG nova.compute.manager [req-e59056bd-b11d-4be3-af47-460307b8cfde req-ed12690f-6569-47f0-835c-1ce09a6a35d7 service nova] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Received event network-vif-deleted-4a3ba85a-80c0-4d5e-a012-7d11194bb806 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 972.967067] env[61243]: INFO nova.compute.manager [req-e59056bd-b11d-4be3-af47-460307b8cfde req-ed12690f-6569-47f0-835c-1ce09a6a35d7 service nova] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Neutron deleted interface 4a3ba85a-80c0-4d5e-a012-7d11194bb806; detaching it from the instance and deleting it from the info cache [ 972.967067] env[61243]: DEBUG nova.network.neutron [req-e59056bd-b11d-4be3-af47-460307b8cfde req-ed12690f-6569-47f0-835c-1ce09a6a35d7 service nova] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.033996] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339251, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.097330] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339258, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.098904] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.099122] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.099329] env[61243]: DEBUG nova.network.neutron [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 973.222051] env[61243]: DEBUG oslo_vmware.api [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339257, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.56375} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.222361] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.222566] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 973.222798] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 973.223014] env[61243]: INFO nova.compute.manager [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Took 2.78 seconds to destroy the instance on the hypervisor. [ 973.223273] env[61243]: DEBUG oslo.service.loopingcall [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.223506] env[61243]: DEBUG nova.compute.manager [-] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 973.224085] env[61243]: DEBUG nova.network.neutron [-] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 973.338910] env[61243]: ERROR nova.scheduler.client.report [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [req-c7fafdcb-4a00-4ce7-8a35-bd5ea7344556] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c7fafdcb-4a00-4ce7-8a35-bd5ea7344556"}]} [ 973.357024] env[61243]: DEBUG nova.scheduler.client.report [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 973.376007] env[61243]: DEBUG nova.scheduler.client.report [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 973.376270] env[61243]: DEBUG nova.compute.provider_tree [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 973.379667] env[61243]: DEBUG nova.compute.manager [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 973.389103] env[61243]: DEBUG nova.scheduler.client.report [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 973.421078] env[61243]: DEBUG nova.scheduler.client.report [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 973.423814] env[61243]: DEBUG nova.network.neutron [-] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.470474] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e2a1626-4936-4df9-badd-8c7e3d13b3ad {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.481372] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe746ede-e84d-4d9e-aa94-831074ddbf00 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.518310] env[61243]: DEBUG nova.compute.manager [req-e59056bd-b11d-4be3-af47-460307b8cfde req-ed12690f-6569-47f0-835c-1ce09a6a35d7 service nova] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Detach interface failed, port_id=4a3ba85a-80c0-4d5e-a012-7d11194bb806, reason: Instance db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 973.535450] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339251, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.877627} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.538141] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e/0c11fc4b-6bf1-4f2f-99f8-bc71dba8db3e.vmdk to [datastore2] 86179fb0-99df-4b10-a815-c19168e9521e/86179fb0-99df-4b10-a815-c19168e9521e.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 973.539181] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-724ca16a-2b34-4a74-878e-fb32573c2c89 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.564973] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 86179fb0-99df-4b10-a815-c19168e9521e/86179fb0-99df-4b10-a815-c19168e9521e.vmdk or device None with type streamOptimized {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 973.568331] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83dd2ca3-b3ca-49e9-bf4a-026eeec91ebf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.592082] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 973.592082] env[61243]: value = "task-1339259" [ 973.592082] env[61243]: _type = "Task" [ 973.592082] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.601240] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339258, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.608336] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339259, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.726091] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61734135-87ec-4150-a810-54395aba908b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.735014] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-810bb851-8e96-489d-bed7-10eca0355863 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.772364] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4fdc010-0b03-4a3d-b1b9-265636002773 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.781138] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec142544-a41b-4f7f-bc3f-e873cb999bd5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.796708] env[61243]: DEBUG nova.compute.provider_tree [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 973.888887] env[61243]: DEBUG nova.network.neutron [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating instance_info_cache with network_info: [{"id": "e21d7ddc-a0d3-4026-a937-f392d5e20fa4", "address": "fa:16:3e:56:77:58", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape21d7ddc-a0", "ovs_interfaceid": "e21d7ddc-a0d3-4026-a937-f392d5e20fa4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.908712] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.927093] env[61243]: INFO nova.compute.manager [-] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Took 1.26 seconds to deallocate network for instance. [ 974.011047] env[61243]: DEBUG nova.network.neutron [-] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.099811] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339258, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.001747} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.102954] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 5103569e-a281-4e89-af69-3a070ab17564/5103569e-a281-4e89-af69-3a070ab17564.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 974.103213] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 974.104161] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9f12935f-e5c3-4982-ba84-9edd3f16d620 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.111096] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339259, 'name': ReconfigVM_Task, 'duration_secs': 0.503124} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.112524] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 86179fb0-99df-4b10-a815-c19168e9521e/86179fb0-99df-4b10-a815-c19168e9521e.vmdk or device None with type streamOptimized {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 974.113385] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 974.113385] env[61243]: value = "task-1339260" [ 974.113385] env[61243]: _type = "Task" [ 974.113385] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.113603] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93eec10b-27d7-4ef6-a31e-92d99059b457 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.127756] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339260, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.129344] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 974.129344] env[61243]: value = "task-1339261" [ 974.129344] env[61243]: _type = "Task" [ 974.129344] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.139237] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339261, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.329826] env[61243]: DEBUG nova.scheduler.client.report [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 123 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 974.332097] env[61243]: DEBUG nova.compute.provider_tree [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 123 to 124 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 974.332097] env[61243]: DEBUG nova.compute.provider_tree [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 974.394368] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.434038] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.514197] env[61243]: INFO nova.compute.manager [-] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Took 1.29 seconds to deallocate network for instance. [ 974.625213] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339260, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.222565} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.625579] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 974.626246] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27a5c2f-1bc1-49c4-89d8-dacdd8c7a1fa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.650672] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 5103569e-a281-4e89-af69-3a070ab17564/5103569e-a281-4e89-af69-3a070ab17564.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 974.651295] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eff069d5-666e-4b78-ae25-699e68ace53a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.667967] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339261, 'name': Rename_Task, 'duration_secs': 0.159621} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.668622] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 974.668871] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-597c9d18-1eb9-4ab0-b718-42e68a598cb9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.674815] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 974.674815] env[61243]: value = "task-1339262" [ 974.674815] env[61243]: _type = "Task" [ 974.674815] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.676345] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 974.676345] env[61243]: value = "task-1339263" [ 974.676345] env[61243]: _type = "Task" [ 974.676345] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.687585] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339262, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.690734] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339263, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.836541] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.786s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.839075] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 5.725s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.839325] env[61243]: DEBUG nova.objects.instance [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61243) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 974.853472] env[61243]: INFO nova.scheduler.client.report [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Deleted allocations for instance d46fb6dc-cdd9-4688-b5e7-9efde0f770c9 [ 974.915581] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa98913-7c6a-4d8f-81db-1183f507de7a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.939964] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e1f373-d1c5-40fa-8a34-0806126b3e24 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.945777] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "5f1c246d-2732-4da4-a00b-b92e2990aa67" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.946015] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "5f1c246d-2732-4da4-a00b-b92e2990aa67" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.952024] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating instance '316572e0-c007-42cb-aaf0-3a8cfcaf24aa' progress to 83 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 974.990031] env[61243]: DEBUG nova.compute.manager [req-81336923-ceaf-4b67-806f-7a7b0153d5f0 req-ad949c9e-9b1f-4c35-af43-49a766c28432 service nova] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Received event network-vif-deleted-6cc2bce1-3536-4637-8cdd-4047a0131f66 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 975.021051] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.192074] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339262, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.192353] env[61243]: DEBUG oslo_vmware.api [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339263, 'name': PowerOnVM_Task, 'duration_secs': 0.496145} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.192594] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 975.295787] env[61243]: DEBUG nova.compute.manager [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.296927] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf71403f-c278-49c2-ac2e-bad88d415adc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.360295] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c632b8f3-84f8-44cd-8a1c-0174c74290b9 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d46fb6dc-cdd9-4688-b5e7-9efde0f770c9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.308s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.448971] env[61243]: DEBUG nova.compute.manager [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 975.458928] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f77cac05-1b5f-4f72-874f-bb0432b32db7 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating instance '316572e0-c007-42cb-aaf0-3a8cfcaf24aa' progress to 100 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 975.687016] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339262, 'name': ReconfigVM_Task, 'duration_secs': 0.70448} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.687443] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 5103569e-a281-4e89-af69-3a070ab17564/5103569e-a281-4e89-af69-3a070ab17564.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 975.688287] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1aa8e448-8df2-48fd-bdc0-8618ffa902fa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.695544] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 975.695544] env[61243]: value = "task-1339264" [ 975.695544] env[61243]: _type = "Task" [ 975.695544] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.703610] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339264, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.816613] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7feb2172-677e-4afe-87f8-131ed859473d tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "86179fb0-99df-4b10-a815-c19168e9521e" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 35.103s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.852059] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7eb64e50-8aee-4198-9378-c5fcaf56a037 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.854242] env[61243]: DEBUG oslo_concurrency.lockutils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.463s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.855709] env[61243]: INFO nova.compute.claims [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 975.974628] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.206773] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339264, 'name': Rename_Task, 'duration_secs': 0.347832} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.207131] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 976.207401] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b09ca7d0-568c-43e0-b9ed-32996024c6c0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.214046] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 976.214046] env[61243]: value = "task-1339265" [ 976.214046] env[61243]: _type = "Task" [ 976.214046] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.222132] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339265, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.383330] env[61243]: DEBUG oslo_concurrency.lockutils [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.383599] env[61243]: DEBUG oslo_concurrency.lockutils [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.724559] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339265, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.887837] env[61243]: INFO nova.compute.manager [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Detaching volume de767a49-762d-43e5-bf2b-b9f14e719ead [ 976.927481] env[61243]: INFO nova.virt.block_device [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Attempting to driver detach volume de767a49-762d-43e5-bf2b-b9f14e719ead from mountpoint /dev/sdb [ 976.927761] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Volume detach. Driver type: vmdk {{(pid=61243) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 976.928226] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285724', 'volume_id': 'de767a49-762d-43e5-bf2b-b9f14e719ead', 'name': 'volume-de767a49-762d-43e5-bf2b-b9f14e719ead', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '99ce6ca3-a478-4ebe-bf1b-42459aacd9ba', 'attached_at': '', 'detached_at': '', 'volume_id': 'de767a49-762d-43e5-bf2b-b9f14e719ead', 'serial': 'de767a49-762d-43e5-bf2b-b9f14e719ead'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 976.929361] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac8a3f1-81c8-45d5-b8ab-f52a7e5c6d7f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.959607] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff2565ae-80a6-4239-a6c9-511e540c2a34 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.968195] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-073226dc-4195-449c-b219-52fcd2e51135 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.999613] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd0abeb-cd33-463e-927a-06a558fcc26e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.018044] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] The volume has not been displaced from its original location: [datastore2] volume-de767a49-762d-43e5-bf2b-b9f14e719ead/volume-de767a49-762d-43e5-bf2b-b9f14e719ead.vmdk. No consolidation needed. {{(pid=61243) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 977.023648] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Reconfiguring VM instance instance-0000003d to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 977.026993] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d865f52-9e2b-4c41-a8c8-bfacce04065c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.061649] env[61243]: DEBUG oslo_vmware.api [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 977.061649] env[61243]: value = "task-1339266" [ 977.061649] env[61243]: _type = "Task" [ 977.061649] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.075983] env[61243]: DEBUG oslo_vmware.api [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339266, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.151887] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cb9b296-cfa8-460b-b210-89895d85ef0d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.160218] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee73c05-2da0-4eee-ba65-088e62d0a7bf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.191823] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c9e225-1dcf-4938-b933-9afeb29d1555 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.200431] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be08339-03b0-4dbc-8203-dc68421a6d29 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.214463] env[61243]: DEBUG nova.compute.provider_tree [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.223792] env[61243]: DEBUG oslo_vmware.api [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339265, 'name': PowerOnVM_Task, 'duration_secs': 0.800717} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.224649] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 977.224862] env[61243]: INFO nova.compute.manager [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Took 10.18 seconds to spawn the instance on the hypervisor. [ 977.225067] env[61243]: DEBUG nova.compute.manager [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 977.225806] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-547960c7-5672-4b9f-bd99-72a1a55c9956 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.241282] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-953ac32a-accb-42c9-b04a-85ef0d18e6ae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.247195] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-08be725f-ee9e-448b-a0b1-1518ec89e5aa tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Suspending the VM {{(pid=61243) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 977.247676] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-268bf7a9-e8a0-4310-87b1-5d27843f8764 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.255332] env[61243]: DEBUG oslo_vmware.api [None req-08be725f-ee9e-448b-a0b1-1518ec89e5aa tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 977.255332] env[61243]: value = "task-1339267" [ 977.255332] env[61243]: _type = "Task" [ 977.255332] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.264450] env[61243]: DEBUG oslo_vmware.api [None req-08be725f-ee9e-448b-a0b1-1518ec89e5aa tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339267, 'name': SuspendVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.573433] env[61243]: DEBUG oslo_vmware.api [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339266, 'name': ReconfigVM_Task, 'duration_secs': 0.296624} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.573724] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Reconfigured VM instance instance-0000003d to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 977.579134] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d79311f-cb1d-4b2d-858b-3a514b0c09a6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.597256] env[61243]: DEBUG oslo_vmware.api [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 977.597256] env[61243]: value = "task-1339268" [ 977.597256] env[61243]: _type = "Task" [ 977.597256] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.606452] env[61243]: DEBUG oslo_vmware.api [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339268, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.720095] env[61243]: DEBUG nova.scheduler.client.report [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 977.746975] env[61243]: INFO nova.compute.manager [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Took 22.91 seconds to build instance. [ 977.768422] env[61243]: DEBUG oslo_vmware.api [None req-08be725f-ee9e-448b-a0b1-1518ec89e5aa tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339267, 'name': SuspendVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.820589] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.820964] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.821251] env[61243]: DEBUG nova.compute.manager [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Going to confirm migration 1 {{(pid=61243) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 978.107060] env[61243]: DEBUG oslo_vmware.api [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339268, 'name': ReconfigVM_Task, 'duration_secs': 0.162404} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.107387] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285724', 'volume_id': 'de767a49-762d-43e5-bf2b-b9f14e719ead', 'name': 'volume-de767a49-762d-43e5-bf2b-b9f14e719ead', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '99ce6ca3-a478-4ebe-bf1b-42459aacd9ba', 'attached_at': '', 'detached_at': '', 'volume_id': 'de767a49-762d-43e5-bf2b-b9f14e719ead', 'serial': 'de767a49-762d-43e5-bf2b-b9f14e719ead'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 978.226871] env[61243]: DEBUG oslo_concurrency.lockutils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.227434] env[61243]: DEBUG nova.compute.manager [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 978.230726] env[61243]: DEBUG oslo_concurrency.lockutils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.207s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.232151] env[61243]: INFO nova.compute.claims [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 978.248010] env[61243]: DEBUG oslo_concurrency.lockutils [None req-adf63d80-11fc-43a6-9b23-4d283494ebb6 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "5103569e-a281-4e89-af69-3a070ab17564" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.419s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.266489] env[61243]: DEBUG oslo_vmware.api [None req-08be725f-ee9e-448b-a0b1-1518ec89e5aa tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339267, 'name': SuspendVM_Task, 'duration_secs': 0.645664} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.267407] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-08be725f-ee9e-448b-a0b1-1518ec89e5aa tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Suspended the VM {{(pid=61243) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 978.267595] env[61243]: DEBUG nova.compute.manager [None req-08be725f-ee9e-448b-a0b1-1518ec89e5aa tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 978.268412] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9acd9054-aab8-4e96-ac85-0d768cf8c1fd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.405553] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.405880] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.406023] env[61243]: DEBUG nova.network.neutron [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 978.406259] env[61243]: DEBUG nova.objects.instance [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lazy-loading 'info_cache' on Instance uuid 316572e0-c007-42cb-aaf0-3a8cfcaf24aa {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.627057] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1d0b9969-5252-4dc3-9bbb-ffe92324ba0d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "interface-b0b5b215-d7ff-4960-9d39-cf792d8d4656-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.627057] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1d0b9969-5252-4dc3-9bbb-ffe92324ba0d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-b0b5b215-d7ff-4960-9d39-cf792d8d4656-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.627057] env[61243]: DEBUG nova.objects.instance [None req-1d0b9969-5252-4dc3-9bbb-ffe92324ba0d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lazy-loading 'flavor' on Instance uuid b0b5b215-d7ff-4960-9d39-cf792d8d4656 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.651240] env[61243]: DEBUG nova.objects.instance [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lazy-loading 'flavor' on Instance uuid 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.736773] env[61243]: DEBUG nova.compute.utils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 978.740241] env[61243]: DEBUG nova.compute.manager [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 978.740241] env[61243]: DEBUG nova.network.neutron [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 978.777815] env[61243]: DEBUG nova.policy [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a96a8e8565f84ba1b2c2406e6a0f224b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8549626a07e941e1ab03452355a0f116', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 978.840790] env[61243]: DEBUG nova.compute.manager [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 978.841859] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76c18dcb-c1de-4f90-98b8-511f2f606d4f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.029719] env[61243]: DEBUG nova.network.neutron [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Successfully created port: b2f45b07-0ee1-45ba-a96b-9e8d62f5159d {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 979.131118] env[61243]: DEBUG nova.objects.instance [None req-1d0b9969-5252-4dc3-9bbb-ffe92324ba0d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lazy-loading 'pci_requests' on Instance uuid b0b5b215-d7ff-4960-9d39-cf792d8d4656 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.240666] env[61243]: DEBUG nova.compute.manager [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 979.353942] env[61243]: INFO nova.compute.manager [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] instance snapshotting [ 979.356246] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe5c509b-2a9f-44f3-9f25-a186b26e7714 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.378149] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5213a9bf-53d4-48e7-9ec5-797dc71a471c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.450365] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749aba6b-d47b-461a-a3bd-a4cd24be7add {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.458056] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d013e46-b21e-495c-af7f-cbbbe30775c9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.489781] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd76fdc5-a60a-4529-b31b-48eb301263f3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.497587] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1176e135-c11c-400f-80eb-5cffe62ebeb5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.511232] env[61243]: DEBUG nova.compute.provider_tree [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.635695] env[61243]: DEBUG nova.objects.base [None req-1d0b9969-5252-4dc3-9bbb-ffe92324ba0d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 979.635695] env[61243]: DEBUG nova.network.neutron [None req-1d0b9969-5252-4dc3-9bbb-ffe92324ba0d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 979.659579] env[61243]: DEBUG oslo_concurrency.lockutils [None req-723fe036-af5c-43a9-8ee1-14f229c3e499 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.276s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.667899] env[61243]: DEBUG nova.network.neutron [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating instance_info_cache with network_info: [{"id": "e21d7ddc-a0d3-4026-a937-f392d5e20fa4", "address": "fa:16:3e:56:77:58", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape21d7ddc-a0", "ovs_interfaceid": "e21d7ddc-a0d3-4026-a937-f392d5e20fa4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.694261] env[61243]: INFO nova.compute.manager [None req-e0ab2092-9103-44f1-9895-88c9b349bd7a tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Resuming [ 979.694834] env[61243]: DEBUG nova.objects.instance [None req-e0ab2092-9103-44f1-9895-88c9b349bd7a tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lazy-loading 'flavor' on Instance uuid 86179fb0-99df-4b10-a815-c19168e9521e {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.729626] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1d0b9969-5252-4dc3-9bbb-ffe92324ba0d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-b0b5b215-d7ff-4960-9d39-cf792d8d4656-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.103s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.893023] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Creating Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 979.893023] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-725bcad9-44d2-4cb0-a1f1-666d327aa910 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.900438] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 979.900438] env[61243]: value = "task-1339269" [ 979.900438] env[61243]: _type = "Task" [ 979.900438] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.909511] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339269, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.014811] env[61243]: DEBUG nova.scheduler.client.report [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.171587] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.171587] env[61243]: DEBUG nova.objects.instance [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lazy-loading 'migration_context' on Instance uuid 316572e0-c007-42cb-aaf0-3a8cfcaf24aa {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 980.252763] env[61243]: DEBUG nova.compute.manager [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 980.278408] env[61243]: DEBUG nova.virt.hardware [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 980.278744] env[61243]: DEBUG nova.virt.hardware [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 980.278917] env[61243]: DEBUG nova.virt.hardware [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 980.279220] env[61243]: DEBUG nova.virt.hardware [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 980.279353] env[61243]: DEBUG nova.virt.hardware [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 980.279549] env[61243]: DEBUG nova.virt.hardware [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 980.279808] env[61243]: DEBUG nova.virt.hardware [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 980.280050] env[61243]: DEBUG nova.virt.hardware [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 980.280315] env[61243]: DEBUG nova.virt.hardware [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 980.280557] env[61243]: DEBUG nova.virt.hardware [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 980.280788] env[61243]: DEBUG nova.virt.hardware [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 980.282095] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b931f8e3-3a23-4e6b-a46f-cf5a425c3e9f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.291359] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ed3cf9-1693-4dec-a626-204da86dbc61 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.412110] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339269, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.520555] env[61243]: DEBUG oslo_concurrency.lockutils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.290s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.521490] env[61243]: DEBUG nova.compute.manager [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 980.525515] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 8.371s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.531568] env[61243]: DEBUG nova.network.neutron [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Successfully updated port: b2f45b07-0ee1-45ba-a96b-9e8d62f5159d {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 980.593530] env[61243]: DEBUG nova.compute.manager [req-a95b114e-1858-4007-ba89-767ff1324ff0 req-254462b4-6ecd-4f50-8b8c-2d663d331702 service nova] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Received event network-vif-plugged-b2f45b07-0ee1-45ba-a96b-9e8d62f5159d {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.594391] env[61243]: DEBUG oslo_concurrency.lockutils [req-a95b114e-1858-4007-ba89-767ff1324ff0 req-254462b4-6ecd-4f50-8b8c-2d663d331702 service nova] Acquiring lock "55f24903-b0b4-4124-afde-f31b2680a53a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.594391] env[61243]: DEBUG oslo_concurrency.lockutils [req-a95b114e-1858-4007-ba89-767ff1324ff0 req-254462b4-6ecd-4f50-8b8c-2d663d331702 service nova] Lock "55f24903-b0b4-4124-afde-f31b2680a53a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.594391] env[61243]: DEBUG oslo_concurrency.lockutils [req-a95b114e-1858-4007-ba89-767ff1324ff0 req-254462b4-6ecd-4f50-8b8c-2d663d331702 service nova] Lock "55f24903-b0b4-4124-afde-f31b2680a53a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.594391] env[61243]: DEBUG nova.compute.manager [req-a95b114e-1858-4007-ba89-767ff1324ff0 req-254462b4-6ecd-4f50-8b8c-2d663d331702 service nova] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] No waiting events found dispatching network-vif-plugged-b2f45b07-0ee1-45ba-a96b-9e8d62f5159d {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 980.594635] env[61243]: WARNING nova.compute.manager [req-a95b114e-1858-4007-ba89-767ff1324ff0 req-254462b4-6ecd-4f50-8b8c-2d663d331702 service nova] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Received unexpected event network-vif-plugged-b2f45b07-0ee1-45ba-a96b-9e8d62f5159d for instance with vm_state building and task_state spawning. [ 980.674297] env[61243]: DEBUG nova.objects.base [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Object Instance<316572e0-c007-42cb-aaf0-3a8cfcaf24aa> lazy-loaded attributes: info_cache,migration_context {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 980.675264] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0a840d6-aba2-421a-a6a1-41779402bfee {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.697500] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-819b3761-6ef9-4bde-94fd-ce4606290dde {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.704533] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e0ab2092-9103-44f1-9895-88c9b349bd7a tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.704723] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e0ab2092-9103-44f1-9895-88c9b349bd7a tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquired lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.705133] env[61243]: DEBUG nova.network.neutron [None req-e0ab2092-9103-44f1-9895-88c9b349bd7a tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 980.706578] env[61243]: DEBUG oslo_vmware.api [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 980.706578] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5260564d-b57f-8941-4363-64cd4584e9ca" [ 980.706578] env[61243]: _type = "Task" [ 980.706578] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.717497] env[61243]: DEBUG oslo_vmware.api [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5260564d-b57f-8941-4363-64cd4584e9ca, 'name': SearchDatastore_Task, 'duration_secs': 0.008643} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.718477] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.767331] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.767331] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.767555] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.767702] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.767881] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.770063] env[61243]: INFO nova.compute.manager [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Terminating instance [ 980.771862] env[61243]: DEBUG nova.compute.manager [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 980.772094] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 980.772918] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b9db64-c52a-42b9-b1e7-62906b116194 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.781402] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 980.781639] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7bae4c85-151f-4c14-9583-0e26445e8e9c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.789492] env[61243]: DEBUG oslo_vmware.api [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 980.789492] env[61243]: value = "task-1339270" [ 980.789492] env[61243]: _type = "Task" [ 980.789492] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.798426] env[61243]: DEBUG oslo_vmware.api [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339270, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.912402] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339269, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.030067] env[61243]: DEBUG nova.compute.utils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 981.039691] env[61243]: DEBUG nova.compute.manager [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 981.039691] env[61243]: DEBUG nova.network.neutron [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 981.040125] env[61243]: DEBUG oslo_concurrency.lockutils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "refresh_cache-55f24903-b0b4-4124-afde-f31b2680a53a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.040295] env[61243]: DEBUG oslo_concurrency.lockutils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquired lock "refresh_cache-55f24903-b0b4-4124-afde-f31b2680a53a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.040477] env[61243]: DEBUG nova.network.neutron [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 981.081760] env[61243]: DEBUG nova.policy [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '785fd146024849efa0b67b1a7f731393', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9db958446334496bab832b95f2363f2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 981.299641] env[61243]: DEBUG oslo_vmware.api [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339270, 'name': PowerOffVM_Task, 'duration_secs': 0.285438} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.299923] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 981.300114] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 981.300375] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f3d317a6-cf88-4e0d-bacb-1ec922edab4c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.387598] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 981.387909] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 981.388181] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Deleting the datastore file [datastore1] 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.388935] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d767a8d0-75cd-4eb5-be36-b02599c53f1d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.395698] env[61243]: DEBUG oslo_vmware.api [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 981.395698] env[61243]: value = "task-1339272" [ 981.395698] env[61243]: _type = "Task" [ 981.395698] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.404230] env[61243]: DEBUG oslo_vmware.api [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339272, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.412102] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339269, 'name': CreateSnapshot_Task, 'duration_secs': 1.106478} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.412391] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Created Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 981.413197] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6bba6ca-5d01-4ea5-a34c-0f845f85156a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.538197] env[61243]: DEBUG nova.compute.manager [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 981.545765] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Applying migration context for instance 316572e0-c007-42cb-aaf0-3a8cfcaf24aa as it has an incoming, in-progress migration 172672ef-2a6b-494c-85d6-065c551cbfca. Migration status is confirming {{(pid=61243) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 981.547205] env[61243]: INFO nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating resource usage from migration 172672ef-2a6b-494c-85d6-065c551cbfca [ 981.583196] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 981.583353] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 3f839aa5-2b9a-4807-b63b-931f74455532 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 981.583465] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance b0b5b215-d7ff-4960-9d39-cf792d8d4656 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 981.583582] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance c1a0d242-4eae-4c03-8341-840b41341f17 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 981.583695] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 86179fb0-99df-4b10-a815-c19168e9521e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 981.583838] env[61243]: WARNING nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 7ce043e6-abdf-4708-80aa-221870375c23 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 981.583966] env[61243]: WARNING nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 981.584091] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 7ca74bdc-c272-4b6e-a4dd-56d97144b197 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 981.584204] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Migration 172672ef-2a6b-494c-85d6-065c551cbfca is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 981.584315] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 316572e0-c007-42cb-aaf0-3a8cfcaf24aa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 981.584447] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 5103569e-a281-4e89-af69-3a070ab17564 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 981.584555] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 55f24903-b0b4-4124-afde-f31b2680a53a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 981.584661] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 5660de64-7152-4832-960e-2deb1c0bc37e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 981.617667] env[61243]: DEBUG nova.network.neutron [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 981.684599] env[61243]: DEBUG oslo_concurrency.lockutils [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "interface-b0b5b215-d7ff-4960-9d39-cf792d8d4656-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.684870] env[61243]: DEBUG oslo_concurrency.lockutils [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-b0b5b215-d7ff-4960-9d39-cf792d8d4656-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.685208] env[61243]: DEBUG nova.objects.instance [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lazy-loading 'flavor' on Instance uuid b0b5b215-d7ff-4960-9d39-cf792d8d4656 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.768193] env[61243]: DEBUG nova.network.neutron [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Successfully created port: 2b065192-815c-492f-8f79-bfa9553ff080 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 981.828106] env[61243]: DEBUG nova.network.neutron [None req-e0ab2092-9103-44f1-9895-88c9b349bd7a tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Updating instance_info_cache with network_info: [{"id": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "address": "fa:16:3e:b0:f1:66", "network": {"id": "49be247d-97e7-4085-8272-4ccf41767d10", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1335943921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "399030c8863346ea97f6da669ff3868b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2248ec9-02", "ovs_interfaceid": "f2248ec9-0230-4aa8-a016-a30ddf4f45f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.905983] env[61243]: DEBUG oslo_vmware.api [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339272, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190813} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.906964] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 981.906964] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 981.906964] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 981.912050] env[61243]: INFO nova.compute.manager [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Took 1.13 seconds to destroy the instance on the hypervisor. [ 981.912050] env[61243]: DEBUG oslo.service.loopingcall [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.912050] env[61243]: DEBUG nova.compute.manager [-] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 981.912050] env[61243]: DEBUG nova.network.neutron [-] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 981.912050] env[61243]: DEBUG nova.network.neutron [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Updating instance_info_cache with network_info: [{"id": "b2f45b07-0ee1-45ba-a96b-9e8d62f5159d", "address": "fa:16:3e:e8:a7:8b", "network": {"id": "c6d4943d-4a8d-41b4-8023-c4f1549d312d", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1045259606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8549626a07e941e1ab03452355a0f116", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2f45b07-0e", "ovs_interfaceid": "b2f45b07-0ee1-45ba-a96b-9e8d62f5159d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.932793] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Creating linked-clone VM from snapshot {{(pid=61243) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 981.933392] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-056c2e04-6778-450e-a61f-4652c1426fba {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.943572] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 981.943572] env[61243]: value = "task-1339273" [ 981.943572] env[61243]: _type = "Task" [ 981.943572] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.952928] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339273, 'name': CloneVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.088079] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 818476f0-e928-44d5-b3d4-101b7a05c9aa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 982.329716] env[61243]: DEBUG nova.objects.instance [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lazy-loading 'pci_requests' on Instance uuid b0b5b215-d7ff-4960-9d39-cf792d8d4656 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.331094] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e0ab2092-9103-44f1-9895-88c9b349bd7a tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Releasing lock "refresh_cache-86179fb0-99df-4b10-a815-c19168e9521e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.333952] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d73cad-5521-410f-bc8f-bf4009596922 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.341779] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e0ab2092-9103-44f1-9895-88c9b349bd7a tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Resuming the VM {{(pid=61243) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 982.343139] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b13a65d9-fdfe-43a4-a3f8-e79b62f6c9bd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.351494] env[61243]: DEBUG oslo_vmware.api [None req-e0ab2092-9103-44f1-9895-88c9b349bd7a tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 982.351494] env[61243]: value = "task-1339274" [ 982.351494] env[61243]: _type = "Task" [ 982.351494] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.360799] env[61243]: DEBUG oslo_vmware.api [None req-e0ab2092-9103-44f1-9895-88c9b349bd7a tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339274, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.414019] env[61243]: DEBUG oslo_concurrency.lockutils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Releasing lock "refresh_cache-55f24903-b0b4-4124-afde-f31b2680a53a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.414444] env[61243]: DEBUG nova.compute.manager [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Instance network_info: |[{"id": "b2f45b07-0ee1-45ba-a96b-9e8d62f5159d", "address": "fa:16:3e:e8:a7:8b", "network": {"id": "c6d4943d-4a8d-41b4-8023-c4f1549d312d", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1045259606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8549626a07e941e1ab03452355a0f116", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2f45b07-0e", "ovs_interfaceid": "b2f45b07-0ee1-45ba-a96b-9e8d62f5159d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 982.414876] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:a7:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24ec44b7-0acf-4ff9-8bb3-4641b74af7a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b2f45b07-0ee1-45ba-a96b-9e8d62f5159d', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 982.423430] env[61243]: DEBUG oslo.service.loopingcall [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.423714] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 982.423957] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3602413f-03f7-47a4-97e7-fafb2a778945 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.445882] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 982.445882] env[61243]: value = "task-1339275" [ 982.445882] env[61243]: _type = "Task" [ 982.445882] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.458170] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339273, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.461198] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339275, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.550752] env[61243]: DEBUG nova.compute.manager [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 982.585298] env[61243]: DEBUG nova.virt.hardware [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 982.585594] env[61243]: DEBUG nova.virt.hardware [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 982.585761] env[61243]: DEBUG nova.virt.hardware [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 982.585950] env[61243]: DEBUG nova.virt.hardware [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 982.586202] env[61243]: DEBUG nova.virt.hardware [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 982.586374] env[61243]: DEBUG nova.virt.hardware [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 982.586593] env[61243]: DEBUG nova.virt.hardware [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 982.587117] env[61243]: DEBUG nova.virt.hardware [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 982.587117] env[61243]: DEBUG nova.virt.hardware [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 982.587117] env[61243]: DEBUG nova.virt.hardware [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 982.587466] env[61243]: DEBUG nova.virt.hardware [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 982.588316] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9230801e-b000-4330-b2fe-44c030a92efa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.591986] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 5f1c246d-2732-4da4-a00b-b92e2990aa67 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 982.592262] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 982.592410] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 982.601454] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aff67dc-3014-42fe-af7e-210b02dcbfb3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.625229] env[61243]: DEBUG nova.compute.manager [req-ba1c3602-940e-4cfb-a669-48dad6943a44 req-f88e81e3-4c34-43af-8ce4-c4ab653e72c9 service nova] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Received event network-changed-b2f45b07-0ee1-45ba-a96b-9e8d62f5159d {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.625472] env[61243]: DEBUG nova.compute.manager [req-ba1c3602-940e-4cfb-a669-48dad6943a44 req-f88e81e3-4c34-43af-8ce4-c4ab653e72c9 service nova] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Refreshing instance network info cache due to event network-changed-b2f45b07-0ee1-45ba-a96b-9e8d62f5159d. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 982.625738] env[61243]: DEBUG oslo_concurrency.lockutils [req-ba1c3602-940e-4cfb-a669-48dad6943a44 req-f88e81e3-4c34-43af-8ce4-c4ab653e72c9 service nova] Acquiring lock "refresh_cache-55f24903-b0b4-4124-afde-f31b2680a53a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.625894] env[61243]: DEBUG oslo_concurrency.lockutils [req-ba1c3602-940e-4cfb-a669-48dad6943a44 req-f88e81e3-4c34-43af-8ce4-c4ab653e72c9 service nova] Acquired lock "refresh_cache-55f24903-b0b4-4124-afde-f31b2680a53a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.626272] env[61243]: DEBUG nova.network.neutron [req-ba1c3602-940e-4cfb-a669-48dad6943a44 req-f88e81e3-4c34-43af-8ce4-c4ab653e72c9 service nova] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Refreshing network info cache for port b2f45b07-0ee1-45ba-a96b-9e8d62f5159d {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 982.830474] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6520d75d-bf6f-44a9-968b-f6187c994ef3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.833662] env[61243]: DEBUG nova.objects.base [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 982.835959] env[61243]: DEBUG nova.network.neutron [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 982.841980] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04408da-afae-4c29-8696-fe9d69ddde36 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.875834] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bad507e-76a4-4bde-aa80-bad66370c3bc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.886836] env[61243]: DEBUG oslo_vmware.api [None req-e0ab2092-9103-44f1-9895-88c9b349bd7a tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339274, 'name': PowerOnVM_Task, 'duration_secs': 0.479961} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.888440] env[61243]: DEBUG nova.policy [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b32f4788825f4a8e82e028cc10347746', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '683c027e4adc4cc4a6b8747a8c8c0de9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 982.891061] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62f2078f-e74d-4cf9-87a8-96ccbe566044 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.895104] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e0ab2092-9103-44f1-9895-88c9b349bd7a tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Resumed the VM {{(pid=61243) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 982.895348] env[61243]: DEBUG nova.compute.manager [None req-e0ab2092-9103-44f1-9895-88c9b349bd7a tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 982.896184] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-731200c9-0166-4626-bf96-94aa224c9435 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.910703] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.962489] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339273, 'name': CloneVM_Task} progress is 95%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.966488] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339275, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.063588] env[61243]: DEBUG nova.network.neutron [-] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.368497] env[61243]: DEBUG nova.network.neutron [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Successfully created port: b91562c3-ce88-4ab3-8413-a5bfeb82aa48 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 983.420830] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 983.459900] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339273, 'name': CloneVM_Task, 'duration_secs': 1.291688} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.463546] env[61243]: INFO nova.virt.vmwareapi.vmops [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Created linked-clone VM from snapshot [ 983.463802] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339275, 'name': CreateVM_Task, 'duration_secs': 0.585301} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.464772] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7fc166-b52a-46bb-8010-5c6937abb02b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.467419] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 983.467923] env[61243]: DEBUG oslo_concurrency.lockutils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.468112] env[61243]: DEBUG oslo_concurrency.lockutils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.468438] env[61243]: DEBUG oslo_concurrency.lockutils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 983.469601] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4df9b060-34c8-40f7-ab01-99d6c9c84887 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.474010] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Uploading image 0a9b6ae7-73bb-4b66-b1f4-ca29058219a1 {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 983.479600] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 983.479600] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528881ad-52b2-36bd-f221-b5746b98f6df" [ 983.479600] env[61243]: _type = "Task" [ 983.479600] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.488049] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528881ad-52b2-36bd-f221-b5746b98f6df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.488960] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Destroying the VM {{(pid=61243) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 983.491159] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7643b38b-e4a2-4d8b-bb30-ae1337e41245 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.499329] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 983.499329] env[61243]: value = "task-1339276" [ 983.499329] env[61243]: _type = "Task" [ 983.499329] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.508385] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339276, 'name': Destroy_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.566823] env[61243]: INFO nova.compute.manager [-] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Took 1.66 seconds to deallocate network for instance. [ 983.606880] env[61243]: DEBUG nova.network.neutron [req-ba1c3602-940e-4cfb-a669-48dad6943a44 req-f88e81e3-4c34-43af-8ce4-c4ab653e72c9 service nova] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Updated VIF entry in instance network info cache for port b2f45b07-0ee1-45ba-a96b-9e8d62f5159d. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 983.607272] env[61243]: DEBUG nova.network.neutron [req-ba1c3602-940e-4cfb-a669-48dad6943a44 req-f88e81e3-4c34-43af-8ce4-c4ab653e72c9 service nova] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Updating instance_info_cache with network_info: [{"id": "b2f45b07-0ee1-45ba-a96b-9e8d62f5159d", "address": "fa:16:3e:e8:a7:8b", "network": {"id": "c6d4943d-4a8d-41b4-8023-c4f1549d312d", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1045259606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8549626a07e941e1ab03452355a0f116", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2f45b07-0e", "ovs_interfaceid": "b2f45b07-0ee1-45ba-a96b-9e8d62f5159d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.848428] env[61243]: DEBUG nova.network.neutron [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Successfully updated port: 2b065192-815c-492f-8f79-bfa9553ff080 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 983.929052] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61243) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 983.929052] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.404s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.929563] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.021s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.931265] env[61243]: INFO nova.compute.claims [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 983.989629] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528881ad-52b2-36bd-f221-b5746b98f6df, 'name': SearchDatastore_Task, 'duration_secs': 0.013305} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.989952] env[61243]: DEBUG oslo_concurrency.lockutils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.990221] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 983.990452] env[61243]: DEBUG oslo_concurrency.lockutils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.990606] env[61243]: DEBUG oslo_concurrency.lockutils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.991012] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.991133] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2fb102b3-6ec0-4dc5-bc67-602090e15386 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.005365] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 984.005570] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 984.006926] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eeccf4a0-0df3-4d8b-a610-33518b84186a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.011973] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339276, 'name': Destroy_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.014932] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 984.014932] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cbf503-5085-3386-8e8b-651c9a986f66" [ 984.014932] env[61243]: _type = "Task" [ 984.014932] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.022387] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cbf503-5085-3386-8e8b-651c9a986f66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.073185] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.110950] env[61243]: DEBUG oslo_concurrency.lockutils [req-ba1c3602-940e-4cfb-a669-48dad6943a44 req-f88e81e3-4c34-43af-8ce4-c4ab653e72c9 service nova] Releasing lock "refresh_cache-55f24903-b0b4-4124-afde-f31b2680a53a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.111064] env[61243]: DEBUG nova.compute.manager [req-ba1c3602-940e-4cfb-a669-48dad6943a44 req-f88e81e3-4c34-43af-8ce4-c4ab653e72c9 service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Received event network-vif-deleted-b22c18f9-e3a1-4c8b-a26a-17c651ca16b5 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.111211] env[61243]: INFO nova.compute.manager [req-ba1c3602-940e-4cfb-a669-48dad6943a44 req-f88e81e3-4c34-43af-8ce4-c4ab653e72c9 service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Neutron deleted interface b22c18f9-e3a1-4c8b-a26a-17c651ca16b5; detaching it from the instance and deleting it from the info cache [ 984.111434] env[61243]: DEBUG nova.network.neutron [req-ba1c3602-940e-4cfb-a669-48dad6943a44 req-f88e81e3-4c34-43af-8ce4-c4ab653e72c9 service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.349679] env[61243]: DEBUG oslo_concurrency.lockutils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "refresh_cache-5660de64-7152-4832-960e-2deb1c0bc37e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.349830] env[61243]: DEBUG oslo_concurrency.lockutils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "refresh_cache-5660de64-7152-4832-960e-2deb1c0bc37e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.349987] env[61243]: DEBUG nova.network.neutron [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 984.511237] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339276, 'name': Destroy_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.525228] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cbf503-5085-3386-8e8b-651c9a986f66, 'name': SearchDatastore_Task, 'duration_secs': 0.030459} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.525228] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8f61a14-5b57-4517-9a4c-9cabcd6a65a0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.529992] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 984.529992] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5201f845-ae90-eb0d-e57f-5799b28c1635" [ 984.529992] env[61243]: _type = "Task" [ 984.529992] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.537448] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5201f845-ae90-eb0d-e57f-5799b28c1635, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.614058] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a8d2095f-aa3b-4971-9610-29cd4735577b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.624015] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fad2557c-81e1-4cf9-ac42-2f8022a89119 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.653173] env[61243]: DEBUG nova.compute.manager [req-ba1c3602-940e-4cfb-a669-48dad6943a44 req-f88e81e3-4c34-43af-8ce4-c4ab653e72c9 service nova] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Detach interface failed, port_id=b22c18f9-e3a1-4c8b-a26a-17c651ca16b5, reason: Instance 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 984.767158] env[61243]: DEBUG nova.compute.manager [req-2cb90244-ccb2-4e4a-81b3-ceac54f7a994 req-86b1d4a7-652d-4074-97a6-38ea3ec17f22 service nova] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Received event network-vif-plugged-2b065192-815c-492f-8f79-bfa9553ff080 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.767515] env[61243]: DEBUG oslo_concurrency.lockutils [req-2cb90244-ccb2-4e4a-81b3-ceac54f7a994 req-86b1d4a7-652d-4074-97a6-38ea3ec17f22 service nova] Acquiring lock "5660de64-7152-4832-960e-2deb1c0bc37e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.767897] env[61243]: DEBUG oslo_concurrency.lockutils [req-2cb90244-ccb2-4e4a-81b3-ceac54f7a994 req-86b1d4a7-652d-4074-97a6-38ea3ec17f22 service nova] Lock "5660de64-7152-4832-960e-2deb1c0bc37e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.769195] env[61243]: DEBUG oslo_concurrency.lockutils [req-2cb90244-ccb2-4e4a-81b3-ceac54f7a994 req-86b1d4a7-652d-4074-97a6-38ea3ec17f22 service nova] Lock "5660de64-7152-4832-960e-2deb1c0bc37e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.769880] env[61243]: DEBUG nova.compute.manager [req-2cb90244-ccb2-4e4a-81b3-ceac54f7a994 req-86b1d4a7-652d-4074-97a6-38ea3ec17f22 service nova] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] No waiting events found dispatching network-vif-plugged-2b065192-815c-492f-8f79-bfa9553ff080 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 984.770113] env[61243]: WARNING nova.compute.manager [req-2cb90244-ccb2-4e4a-81b3-ceac54f7a994 req-86b1d4a7-652d-4074-97a6-38ea3ec17f22 service nova] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Received unexpected event network-vif-plugged-2b065192-815c-492f-8f79-bfa9553ff080 for instance with vm_state building and task_state spawning. [ 984.771596] env[61243]: DEBUG nova.compute.manager [req-2cb90244-ccb2-4e4a-81b3-ceac54f7a994 req-86b1d4a7-652d-4074-97a6-38ea3ec17f22 service nova] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Received event network-changed-2b065192-815c-492f-8f79-bfa9553ff080 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.771659] env[61243]: DEBUG nova.compute.manager [req-2cb90244-ccb2-4e4a-81b3-ceac54f7a994 req-86b1d4a7-652d-4074-97a6-38ea3ec17f22 service nova] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Refreshing instance network info cache due to event network-changed-2b065192-815c-492f-8f79-bfa9553ff080. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 984.772861] env[61243]: DEBUG oslo_concurrency.lockutils [req-2cb90244-ccb2-4e4a-81b3-ceac54f7a994 req-86b1d4a7-652d-4074-97a6-38ea3ec17f22 service nova] Acquiring lock "refresh_cache-5660de64-7152-4832-960e-2deb1c0bc37e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.896963] env[61243]: DEBUG nova.network.neutron [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 985.010858] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339276, 'name': Destroy_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.041113] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5201f845-ae90-eb0d-e57f-5799b28c1635, 'name': SearchDatastore_Task, 'duration_secs': 0.050324} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.043727] env[61243]: DEBUG oslo_concurrency.lockutils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.044013] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 55f24903-b0b4-4124-afde-f31b2680a53a/55f24903-b0b4-4124-afde-f31b2680a53a.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 985.044509] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88e6736b-0a2e-4ca1-8200-b3bd42bb0a82 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.051554] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 985.051554] env[61243]: value = "task-1339277" [ 985.051554] env[61243]: _type = "Task" [ 985.051554] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.062530] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339277, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.148127] env[61243]: DEBUG nova.network.neutron [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Updating instance_info_cache with network_info: [{"id": "2b065192-815c-492f-8f79-bfa9553ff080", "address": "fa:16:3e:94:3e:9e", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b065192-81", "ovs_interfaceid": "2b065192-815c-492f-8f79-bfa9553ff080", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.172155] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8291e0-f350-4ce3-9bdc-f4e7b69d5536 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.179826] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54da128c-dcdf-4cb3-977b-28e0ed5f64f8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.212533] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d61619d-a8ea-4fbb-bab4-581f102b84d4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.220508] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75668536-205c-49f6-8362-29e873adbe8b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.233945] env[61243]: DEBUG nova.compute.provider_tree [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.238890] env[61243]: DEBUG nova.network.neutron [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Successfully updated port: b91562c3-ce88-4ab3-8413-a5bfeb82aa48 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 985.513420] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339276, 'name': Destroy_Task, 'duration_secs': 1.763937} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.513810] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Destroyed the VM [ 985.514206] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Deleting Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 985.514466] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a10449f8-ae79-4521-ba46-c1c2d33bea53 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.521696] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 985.521696] env[61243]: value = "task-1339278" [ 985.521696] env[61243]: _type = "Task" [ 985.521696] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.530022] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339278, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.561858] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339277, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.650746] env[61243]: DEBUG oslo_concurrency.lockutils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "refresh_cache-5660de64-7152-4832-960e-2deb1c0bc37e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.651263] env[61243]: DEBUG nova.compute.manager [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Instance network_info: |[{"id": "2b065192-815c-492f-8f79-bfa9553ff080", "address": "fa:16:3e:94:3e:9e", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b065192-81", "ovs_interfaceid": "2b065192-815c-492f-8f79-bfa9553ff080", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 985.651547] env[61243]: DEBUG oslo_concurrency.lockutils [req-2cb90244-ccb2-4e4a-81b3-ceac54f7a994 req-86b1d4a7-652d-4074-97a6-38ea3ec17f22 service nova] Acquired lock "refresh_cache-5660de64-7152-4832-960e-2deb1c0bc37e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.651704] env[61243]: DEBUG nova.network.neutron [req-2cb90244-ccb2-4e4a-81b3-ceac54f7a994 req-86b1d4a7-652d-4074-97a6-38ea3ec17f22 service nova] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Refreshing network info cache for port 2b065192-815c-492f-8f79-bfa9553ff080 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 985.653157] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:3e:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a071ecf4-e713-4f97-9271-8c17952f6dee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b065192-815c-492f-8f79-bfa9553ff080', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 985.660830] env[61243]: DEBUG oslo.service.loopingcall [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.661351] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 985.662168] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dd2bd0ef-05f0-4234-8835-7d4278d68113 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.681662] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 985.681662] env[61243]: value = "task-1339279" [ 985.681662] env[61243]: _type = "Task" [ 985.681662] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.690958] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339279, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.738567] env[61243]: DEBUG nova.scheduler.client.report [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 985.743066] env[61243]: DEBUG oslo_concurrency.lockutils [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.743320] env[61243]: DEBUG oslo_concurrency.lockutils [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.743523] env[61243]: DEBUG nova.network.neutron [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.031288] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339278, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.063845] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339277, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.616714} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.064171] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 55f24903-b0b4-4124-afde-f31b2680a53a/55f24903-b0b4-4124-afde-f31b2680a53a.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 986.064402] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 986.064709] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da202b63-dbec-4d3f-96ad-728eb2fa9072 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.071662] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 986.071662] env[61243]: value = "task-1339280" [ 986.071662] env[61243]: _type = "Task" [ 986.071662] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.086461] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339280, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.191459] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339279, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.248219] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.319s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.248931] env[61243]: DEBUG nova.compute.manager [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 986.253519] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.820s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.255037] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.255572] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.235s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.255730] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.257304] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.283s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.258788] env[61243]: INFO nova.compute.claims [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 986.282650] env[61243]: INFO nova.scheduler.client.report [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Deleted allocations for instance db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09 [ 986.284188] env[61243]: INFO nova.scheduler.client.report [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleted allocations for instance 7ce043e6-abdf-4708-80aa-221870375c23 [ 986.310996] env[61243]: WARNING nova.network.neutron [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] 7724b87f-3721-42a5-81ce-27ae98fc99c2 already exists in list: networks containing: ['7724b87f-3721-42a5-81ce-27ae98fc99c2']. ignoring it [ 986.490980] env[61243]: DEBUG nova.network.neutron [req-2cb90244-ccb2-4e4a-81b3-ceac54f7a994 req-86b1d4a7-652d-4074-97a6-38ea3ec17f22 service nova] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Updated VIF entry in instance network info cache for port 2b065192-815c-492f-8f79-bfa9553ff080. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 986.491377] env[61243]: DEBUG nova.network.neutron [req-2cb90244-ccb2-4e4a-81b3-ceac54f7a994 req-86b1d4a7-652d-4074-97a6-38ea3ec17f22 service nova] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Updating instance_info_cache with network_info: [{"id": "2b065192-815c-492f-8f79-bfa9553ff080", "address": "fa:16:3e:94:3e:9e", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b065192-81", "ovs_interfaceid": "2b065192-815c-492f-8f79-bfa9553ff080", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.533673] env[61243]: DEBUG oslo_vmware.api [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339278, 'name': RemoveSnapshot_Task, 'duration_secs': 0.735587} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.533950] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Deleted Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 986.581377] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339280, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066259} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.581651] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 986.582497] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686215ec-07ea-41fa-8347-780fcf2720c5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.607468] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 55f24903-b0b4-4124-afde-f31b2680a53a/55f24903-b0b4-4124-afde-f31b2680a53a.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 986.608059] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4cc754a0-851e-452c-be56-1444cd5a4d87 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.629030] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 986.629030] env[61243]: value = "task-1339281" [ 986.629030] env[61243]: _type = "Task" [ 986.629030] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.636308] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339281, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.656904] env[61243]: DEBUG nova.network.neutron [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Updating instance_info_cache with network_info: [{"id": "2333ab37-7807-45d2-a74e-05404217df6d", "address": "fa:16:3e:ea:93:91", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2333ab37-78", "ovs_interfaceid": "2333ab37-7807-45d2-a74e-05404217df6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b91562c3-ce88-4ab3-8413-a5bfeb82aa48", "address": "fa:16:3e:2b:fe:34", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb91562c3-ce", "ovs_interfaceid": "b91562c3-ce88-4ab3-8413-a5bfeb82aa48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.693899] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339279, 'name': CreateVM_Task, 'duration_secs': 0.525924} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.694089] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 986.694749] env[61243]: DEBUG oslo_concurrency.lockutils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.694926] env[61243]: DEBUG oslo_concurrency.lockutils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.695278] env[61243]: DEBUG oslo_concurrency.lockutils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 986.695540] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87dff6a0-0fa6-476c-8177-8749ff37cb6a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.699770] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 986.699770] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5264dc54-255b-12ef-3bc1-abc169250a0a" [ 986.699770] env[61243]: _type = "Task" [ 986.699770] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.707118] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5264dc54-255b-12ef-3bc1-abc169250a0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.755292] env[61243]: DEBUG nova.compute.utils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 986.756616] env[61243]: DEBUG nova.compute.manager [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 986.756808] env[61243]: DEBUG nova.network.neutron [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 986.796339] env[61243]: DEBUG oslo_concurrency.lockutils [None req-46a4b37e-fac1-40e9-82db-9b4c1661dc8d tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "7ce043e6-abdf-4708-80aa-221870375c23" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.369s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.798713] env[61243]: DEBUG nova.policy [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06e247d66c8d43218302a35ead439405', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e641e72e4d2048218ef403456dbb2b82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 986.800259] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c5432c58-780c-4269-babc-68a22e088cc2 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.339s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.899165] env[61243]: DEBUG nova.compute.manager [req-2a051109-699d-42ac-b840-58cb2febd72b req-f1a193d2-82d4-4f88-88ec-ebe584be4163 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Received event network-vif-plugged-b91562c3-ce88-4ab3-8413-a5bfeb82aa48 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 986.899414] env[61243]: DEBUG oslo_concurrency.lockutils [req-2a051109-699d-42ac-b840-58cb2febd72b req-f1a193d2-82d4-4f88-88ec-ebe584be4163 service nova] Acquiring lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.899618] env[61243]: DEBUG oslo_concurrency.lockutils [req-2a051109-699d-42ac-b840-58cb2febd72b req-f1a193d2-82d4-4f88-88ec-ebe584be4163 service nova] Lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.899811] env[61243]: DEBUG oslo_concurrency.lockutils [req-2a051109-699d-42ac-b840-58cb2febd72b req-f1a193d2-82d4-4f88-88ec-ebe584be4163 service nova] Lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.900153] env[61243]: DEBUG nova.compute.manager [req-2a051109-699d-42ac-b840-58cb2febd72b req-f1a193d2-82d4-4f88-88ec-ebe584be4163 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] No waiting events found dispatching network-vif-plugged-b91562c3-ce88-4ab3-8413-a5bfeb82aa48 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 986.900153] env[61243]: WARNING nova.compute.manager [req-2a051109-699d-42ac-b840-58cb2febd72b req-f1a193d2-82d4-4f88-88ec-ebe584be4163 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Received unexpected event network-vif-plugged-b91562c3-ce88-4ab3-8413-a5bfeb82aa48 for instance with vm_state active and task_state None. [ 986.900312] env[61243]: DEBUG nova.compute.manager [req-2a051109-699d-42ac-b840-58cb2febd72b req-f1a193d2-82d4-4f88-88ec-ebe584be4163 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Received event network-changed-b91562c3-ce88-4ab3-8413-a5bfeb82aa48 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 986.900446] env[61243]: DEBUG nova.compute.manager [req-2a051109-699d-42ac-b840-58cb2febd72b req-f1a193d2-82d4-4f88-88ec-ebe584be4163 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Refreshing instance network info cache due to event network-changed-b91562c3-ce88-4ab3-8413-a5bfeb82aa48. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 986.900618] env[61243]: DEBUG oslo_concurrency.lockutils [req-2a051109-699d-42ac-b840-58cb2febd72b req-f1a193d2-82d4-4f88-88ec-ebe584be4163 service nova] Acquiring lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.994208] env[61243]: DEBUG oslo_concurrency.lockutils [req-2cb90244-ccb2-4e4a-81b3-ceac54f7a994 req-86b1d4a7-652d-4074-97a6-38ea3ec17f22 service nova] Releasing lock "refresh_cache-5660de64-7152-4832-960e-2deb1c0bc37e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.041537] env[61243]: WARNING nova.compute.manager [None req-5d749b95-94c2-4c56-b9b8-0005e07af9e8 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Image not found during snapshot: nova.exception.ImageNotFound: Image 0a9b6ae7-73bb-4b66-b1f4-ca29058219a1 could not be found. [ 987.054869] env[61243]: DEBUG nova.network.neutron [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Successfully created port: 75bf3cd4-a26f-4ca5-ba11-69971cc910c8 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 987.139918] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339281, 'name': ReconfigVM_Task, 'duration_secs': 0.462022} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.140085] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 55f24903-b0b4-4124-afde-f31b2680a53a/55f24903-b0b4-4124-afde-f31b2680a53a.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.140592] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a42fcc51-b597-4b3a-88dc-09d8483cc558 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.147361] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 987.147361] env[61243]: value = "task-1339282" [ 987.147361] env[61243]: _type = "Task" [ 987.147361] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.155315] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339282, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.160529] env[61243]: DEBUG oslo_concurrency.lockutils [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.160529] env[61243]: DEBUG oslo_concurrency.lockutils [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.160529] env[61243]: DEBUG oslo_concurrency.lockutils [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.160529] env[61243]: DEBUG oslo_concurrency.lockutils [req-2a051109-699d-42ac-b840-58cb2febd72b req-f1a193d2-82d4-4f88-88ec-ebe584be4163 service nova] Acquired lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.160529] env[61243]: DEBUG nova.network.neutron [req-2a051109-699d-42ac-b840-58cb2febd72b req-f1a193d2-82d4-4f88-88ec-ebe584be4163 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Refreshing network info cache for port b91562c3-ce88-4ab3-8413-a5bfeb82aa48 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 987.162029] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca44f80-0bdb-423a-8aa9-772e7acce32c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.181327] env[61243]: DEBUG nova.virt.hardware [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.181623] env[61243]: DEBUG nova.virt.hardware [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.181797] env[61243]: DEBUG nova.virt.hardware [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.182049] env[61243]: DEBUG nova.virt.hardware [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.182398] env[61243]: DEBUG nova.virt.hardware [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.182398] env[61243]: DEBUG nova.virt.hardware [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.182612] env[61243]: DEBUG nova.virt.hardware [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.182800] env[61243]: DEBUG nova.virt.hardware [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.182990] env[61243]: DEBUG nova.virt.hardware [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.183184] env[61243]: DEBUG nova.virt.hardware [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.183417] env[61243]: DEBUG nova.virt.hardware [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.190450] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Reconfiguring VM to attach interface {{(pid=61243) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 987.191524] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95e396c7-6512-4d49-9e4c-0b96e5c5e675 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.215189] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5264dc54-255b-12ef-3bc1-abc169250a0a, 'name': SearchDatastore_Task, 'duration_secs': 0.105047} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.217074] env[61243]: DEBUG oslo_concurrency.lockutils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.217074] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 987.217074] env[61243]: DEBUG oslo_concurrency.lockutils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.217322] env[61243]: DEBUG oslo_concurrency.lockutils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.217322] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 987.217619] env[61243]: DEBUG oslo_vmware.api [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 987.217619] env[61243]: value = "task-1339283" [ 987.217619] env[61243]: _type = "Task" [ 987.217619] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.217853] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5f1e6a1f-03a2-4d2f-abac-85cdd1e1cf0f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.229680] env[61243]: DEBUG oslo_vmware.api [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339283, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.230683] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 987.230922] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 987.231758] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39ab95b8-a743-4382-849a-4e3fa7304e76 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.237683] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 987.237683] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528f53af-dac7-aa65-f2e4-dee13c0b85e5" [ 987.237683] env[61243]: _type = "Task" [ 987.237683] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.247024] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528f53af-dac7-aa65-f2e4-dee13c0b85e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.259818] env[61243]: DEBUG nova.compute.manager [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 987.443322] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dee5987-f472-4c03-9a01-f6b721067adc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.451676] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e5cb99-e1b1-4214-866b-fb18d8cb5742 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.482055] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec25b35f-a913-4372-84a8-d777956d74aa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.490697] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1efa09-40bb-4c05-80f7-2859bfd819ce {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.505552] env[61243]: DEBUG nova.compute.provider_tree [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.659033] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339282, 'name': Rename_Task, 'duration_secs': 0.140551} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.659033] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 987.659231] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-764df9cb-3e58-4518-927d-db79583dca63 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.668062] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 987.668062] env[61243]: value = "task-1339284" [ 987.668062] env[61243]: _type = "Task" [ 987.668062] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.677564] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339284, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.730093] env[61243]: DEBUG oslo_vmware.api [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339283, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.750817] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528f53af-dac7-aa65-f2e4-dee13c0b85e5, 'name': SearchDatastore_Task, 'duration_secs': 0.009554} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.751830] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1249316f-782f-429d-978d-ded378ce47b7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.758948] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 987.758948] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5287a9bb-c2f6-754e-2c7a-a6b3be4d14a9" [ 987.758948] env[61243]: _type = "Task" [ 987.758948] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.779185] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5287a9bb-c2f6-754e-2c7a-a6b3be4d14a9, 'name': SearchDatastore_Task, 'duration_secs': 0.008893} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.779461] env[61243]: DEBUG oslo_concurrency.lockutils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.779720] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 5660de64-7152-4832-960e-2deb1c0bc37e/5660de64-7152-4832-960e-2deb1c0bc37e.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 987.779987] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-17b41df2-27ef-4ebe-9a1a-3bc9f26eae00 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.791093] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 987.791093] env[61243]: value = "task-1339285" [ 987.791093] env[61243]: _type = "Task" [ 987.791093] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.803732] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339285, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.900478] env[61243]: DEBUG nova.network.neutron [req-2a051109-699d-42ac-b840-58cb2febd72b req-f1a193d2-82d4-4f88-88ec-ebe584be4163 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Updated VIF entry in instance network info cache for port b91562c3-ce88-4ab3-8413-a5bfeb82aa48. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 987.900930] env[61243]: DEBUG nova.network.neutron [req-2a051109-699d-42ac-b840-58cb2febd72b req-f1a193d2-82d4-4f88-88ec-ebe584be4163 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Updating instance_info_cache with network_info: [{"id": "2333ab37-7807-45d2-a74e-05404217df6d", "address": "fa:16:3e:ea:93:91", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2333ab37-78", "ovs_interfaceid": "2333ab37-7807-45d2-a74e-05404217df6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b91562c3-ce88-4ab3-8413-a5bfeb82aa48", "address": "fa:16:3e:2b:fe:34", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb91562c3-ce", "ovs_interfaceid": "b91562c3-ce88-4ab3-8413-a5bfeb82aa48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.980893] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "2c4217cd-af4a-4d64-a422-73955326df03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.981171] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "2c4217cd-af4a-4d64-a422-73955326df03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.010046] env[61243]: DEBUG nova.scheduler.client.report [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.180368] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339284, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.230975] env[61243]: DEBUG oslo_vmware.api [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339283, 'name': ReconfigVM_Task, 'duration_secs': 0.853096} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.231570] env[61243]: DEBUG oslo_concurrency.lockutils [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.231790] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Reconfigured VM to attach interface {{(pid=61243) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 988.273394] env[61243]: DEBUG nova.compute.manager [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 988.303601] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339285, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.307238] env[61243]: DEBUG nova.virt.hardware [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 988.307238] env[61243]: DEBUG nova.virt.hardware [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 988.307238] env[61243]: DEBUG nova.virt.hardware [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 988.307238] env[61243]: DEBUG nova.virt.hardware [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 988.307238] env[61243]: DEBUG nova.virt.hardware [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 988.307238] env[61243]: DEBUG nova.virt.hardware [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 988.307238] env[61243]: DEBUG nova.virt.hardware [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 988.307730] env[61243]: DEBUG nova.virt.hardware [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 988.307730] env[61243]: DEBUG nova.virt.hardware [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 988.307854] env[61243]: DEBUG nova.virt.hardware [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 988.308025] env[61243]: DEBUG nova.virt.hardware [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 988.308932] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99862698-55ba-438e-b5f7-5e1fff1a3033 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.318641] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18bae2d1-cd0f-4896-9a39-775c85b87ee0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.336041] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "5103569e-a281-4e89-af69-3a070ab17564" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.336351] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "5103569e-a281-4e89-af69-3a070ab17564" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.336615] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "5103569e-a281-4e89-af69-3a070ab17564-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.336843] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "5103569e-a281-4e89-af69-3a070ab17564-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.337044] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "5103569e-a281-4e89-af69-3a070ab17564-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.339270] env[61243]: INFO nova.compute.manager [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Terminating instance [ 988.341258] env[61243]: DEBUG nova.compute.manager [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 988.341481] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 988.342341] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409e6aaa-fb2b-4bb0-a6b5-7590f7b0e6d0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.350270] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 988.350617] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d888e2b7-d024-4f39-8f47-b0adc31c51d6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.358091] env[61243]: DEBUG oslo_vmware.api [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 988.358091] env[61243]: value = "task-1339286" [ 988.358091] env[61243]: _type = "Task" [ 988.358091] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.369289] env[61243]: DEBUG oslo_vmware.api [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339286, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.404600] env[61243]: DEBUG oslo_concurrency.lockutils [req-2a051109-699d-42ac-b840-58cb2febd72b req-f1a193d2-82d4-4f88-88ec-ebe584be4163 service nova] Releasing lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.483760] env[61243]: DEBUG nova.compute.manager [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 988.514964] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.258s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.515822] env[61243]: DEBUG nova.compute.manager [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 988.518735] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 7.800s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.610287] env[61243]: DEBUG nova.network.neutron [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Successfully updated port: 75bf3cd4-a26f-4ca5-ba11-69971cc910c8 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 988.680074] env[61243]: DEBUG oslo_vmware.api [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339284, 'name': PowerOnVM_Task, 'duration_secs': 0.921027} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.680472] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 988.680697] env[61243]: INFO nova.compute.manager [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Took 8.43 seconds to spawn the instance on the hypervisor. [ 988.680881] env[61243]: DEBUG nova.compute.manager [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.681695] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01222df7-dd21-4c80-a762-75608d6e5fa0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.736378] env[61243]: DEBUG oslo_concurrency.lockutils [None req-19b51b97-7d65-4786-8bda-e9067a0a1376 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-b0b5b215-d7ff-4960-9d39-cf792d8d4656-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.051s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.801694] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339285, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523901} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.802047] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 5660de64-7152-4832-960e-2deb1c0bc37e/5660de64-7152-4832-960e-2deb1c0bc37e.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 988.802285] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 988.802583] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad74f0a9-2c4f-4f2a-b1ad-08dc5624ccd0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.809867] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 988.809867] env[61243]: value = "task-1339287" [ 988.809867] env[61243]: _type = "Task" [ 988.809867] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.818503] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339287, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.869190] env[61243]: DEBUG oslo_vmware.api [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339286, 'name': PowerOffVM_Task, 'duration_secs': 0.321887} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.869526] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 988.869709] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 988.869971] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-72cae5e3-e90c-445e-9070-6ea066ff5c9a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.924079] env[61243]: DEBUG nova.compute.manager [req-95d39951-5da0-4ee3-8079-045d2c28d9ce req-36f22a37-7346-4691-bc82-6877d22d53a4 service nova] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Received event network-vif-plugged-75bf3cd4-a26f-4ca5-ba11-69971cc910c8 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.924304] env[61243]: DEBUG oslo_concurrency.lockutils [req-95d39951-5da0-4ee3-8079-045d2c28d9ce req-36f22a37-7346-4691-bc82-6877d22d53a4 service nova] Acquiring lock "818476f0-e928-44d5-b3d4-101b7a05c9aa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.924518] env[61243]: DEBUG oslo_concurrency.lockutils [req-95d39951-5da0-4ee3-8079-045d2c28d9ce req-36f22a37-7346-4691-bc82-6877d22d53a4 service nova] Lock "818476f0-e928-44d5-b3d4-101b7a05c9aa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.924692] env[61243]: DEBUG oslo_concurrency.lockutils [req-95d39951-5da0-4ee3-8079-045d2c28d9ce req-36f22a37-7346-4691-bc82-6877d22d53a4 service nova] Lock "818476f0-e928-44d5-b3d4-101b7a05c9aa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.924865] env[61243]: DEBUG nova.compute.manager [req-95d39951-5da0-4ee3-8079-045d2c28d9ce req-36f22a37-7346-4691-bc82-6877d22d53a4 service nova] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] No waiting events found dispatching network-vif-plugged-75bf3cd4-a26f-4ca5-ba11-69971cc910c8 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 988.925049] env[61243]: WARNING nova.compute.manager [req-95d39951-5da0-4ee3-8079-045d2c28d9ce req-36f22a37-7346-4691-bc82-6877d22d53a4 service nova] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Received unexpected event network-vif-plugged-75bf3cd4-a26f-4ca5-ba11-69971cc910c8 for instance with vm_state building and task_state spawning. [ 988.925222] env[61243]: DEBUG nova.compute.manager [req-95d39951-5da0-4ee3-8079-045d2c28d9ce req-36f22a37-7346-4691-bc82-6877d22d53a4 service nova] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Received event network-changed-75bf3cd4-a26f-4ca5-ba11-69971cc910c8 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.925383] env[61243]: DEBUG nova.compute.manager [req-95d39951-5da0-4ee3-8079-045d2c28d9ce req-36f22a37-7346-4691-bc82-6877d22d53a4 service nova] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Refreshing instance network info cache due to event network-changed-75bf3cd4-a26f-4ca5-ba11-69971cc910c8. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 988.925574] env[61243]: DEBUG oslo_concurrency.lockutils [req-95d39951-5da0-4ee3-8079-045d2c28d9ce req-36f22a37-7346-4691-bc82-6877d22d53a4 service nova] Acquiring lock "refresh_cache-818476f0-e928-44d5-b3d4-101b7a05c9aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.925715] env[61243]: DEBUG oslo_concurrency.lockutils [req-95d39951-5da0-4ee3-8079-045d2c28d9ce req-36f22a37-7346-4691-bc82-6877d22d53a4 service nova] Acquired lock "refresh_cache-818476f0-e928-44d5-b3d4-101b7a05c9aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.925874] env[61243]: DEBUG nova.network.neutron [req-95d39951-5da0-4ee3-8079-045d2c28d9ce req-36f22a37-7346-4691-bc82-6877d22d53a4 service nova] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Refreshing network info cache for port 75bf3cd4-a26f-4ca5-ba11-69971cc910c8 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 988.978339] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 988.978609] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 988.978766] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Deleting the datastore file [datastore2] 5103569e-a281-4e89-af69-3a070ab17564 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 988.979063] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-89e50478-0be5-4292-9d01-17082220a8bb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.986619] env[61243]: DEBUG oslo_vmware.api [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for the task: (returnval){ [ 988.986619] env[61243]: value = "task-1339289" [ 988.986619] env[61243]: _type = "Task" [ 988.986619] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.004166] env[61243]: DEBUG oslo_vmware.api [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339289, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.013318] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.026268] env[61243]: DEBUG nova.compute.utils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 989.028627] env[61243]: DEBUG nova.compute.manager [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 989.028627] env[61243]: DEBUG nova.network.neutron [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 989.071797] env[61243]: DEBUG nova.policy [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b9748805e2c48e4a8fd0354e916aa0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '324ed556b6f5419fa0376a7cfc508827', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 989.112882] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "refresh_cache-818476f0-e928-44d5-b3d4-101b7a05c9aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.200861] env[61243]: INFO nova.compute.manager [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Took 17.83 seconds to build instance. [ 989.214068] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-777c9caf-cbc0-4567-b986-30c0dba0aaf2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.223516] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f51b7f9-d132-4d0e-953f-897897e926bc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.257977] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-522f3230-f8e0-46cf-a2cb-3c8ba5357cb4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.267645] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac2da27-57ce-4b54-a40f-024d2def30c1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.282181] env[61243]: DEBUG nova.compute.provider_tree [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.319187] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339287, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066255} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.319479] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 989.320142] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1040342-5d1d-48fb-a48b-48c99e728654 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.342486] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 5660de64-7152-4832-960e-2deb1c0bc37e/5660de64-7152-4832-960e-2deb1c0bc37e.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 989.342829] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23edfbe5-bf62-4618-90d0-d8fc7d8b0460 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.363818] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 989.363818] env[61243]: value = "task-1339290" [ 989.363818] env[61243]: _type = "Task" [ 989.363818] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.373078] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339290, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.496763] env[61243]: DEBUG oslo_vmware.api [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339289, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.497622] env[61243]: DEBUG nova.network.neutron [req-95d39951-5da0-4ee3-8079-045d2c28d9ce req-36f22a37-7346-4691-bc82-6877d22d53a4 service nova] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 989.532280] env[61243]: DEBUG nova.compute.manager [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 989.632259] env[61243]: DEBUG nova.network.neutron [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Successfully created port: 6fddda84-0767-4709-ac83-fc4e6060da3c {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 989.702557] env[61243]: DEBUG nova.network.neutron [req-95d39951-5da0-4ee3-8079-045d2c28d9ce req-36f22a37-7346-4691-bc82-6877d22d53a4 service nova] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.704030] env[61243]: DEBUG oslo_concurrency.lockutils [None req-104e7813-da0e-4b32-b83e-57e1215431d3 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.339s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.784580] env[61243]: DEBUG nova.scheduler.client.report [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 989.874939] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339290, 'name': ReconfigVM_Task, 'duration_secs': 0.392222} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.878467] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 5660de64-7152-4832-960e-2deb1c0bc37e/5660de64-7152-4832-960e-2deb1c0bc37e.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.879208] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5baa30b5-c281-4728-a947-39eca9b06794 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.887700] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 989.887700] env[61243]: value = "task-1339291" [ 989.887700] env[61243]: _type = "Task" [ 989.887700] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.897057] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339291, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.997482] env[61243]: DEBUG oslo_vmware.api [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Task: {'id': task-1339289, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.529976} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.997748] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 989.997943] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 989.998135] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 989.998315] env[61243]: INFO nova.compute.manager [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Took 1.66 seconds to destroy the instance on the hypervisor. [ 989.998616] env[61243]: DEBUG oslo.service.loopingcall [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.998817] env[61243]: DEBUG nova.compute.manager [-] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 989.998919] env[61243]: DEBUG nova.network.neutron [-] [instance: 5103569e-a281-4e89-af69-3a070ab17564] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 990.207795] env[61243]: DEBUG oslo_concurrency.lockutils [req-95d39951-5da0-4ee3-8079-045d2c28d9ce req-36f22a37-7346-4691-bc82-6877d22d53a4 service nova] Releasing lock "refresh_cache-818476f0-e928-44d5-b3d4-101b7a05c9aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.207795] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "refresh_cache-818476f0-e928-44d5-b3d4-101b7a05c9aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.207795] env[61243]: DEBUG nova.network.neutron [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 990.266805] env[61243]: DEBUG oslo_concurrency.lockutils [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "86179fb0-99df-4b10-a815-c19168e9521e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.271020] env[61243]: DEBUG oslo_concurrency.lockutils [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "86179fb0-99df-4b10-a815-c19168e9521e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.271020] env[61243]: DEBUG oslo_concurrency.lockutils [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "86179fb0-99df-4b10-a815-c19168e9521e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.271020] env[61243]: DEBUG oslo_concurrency.lockutils [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "86179fb0-99df-4b10-a815-c19168e9521e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.271020] env[61243]: DEBUG oslo_concurrency.lockutils [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "86179fb0-99df-4b10-a815-c19168e9521e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.271020] env[61243]: INFO nova.compute.manager [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Terminating instance [ 990.273416] env[61243]: DEBUG nova.compute.manager [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 990.273646] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 990.274515] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f849fa-04a4-40d6-a3ab-289678019918 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.288781] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 990.291884] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-efb2437b-18c5-458c-ab5c-6bf32456f2dd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.308252] env[61243]: DEBUG oslo_vmware.api [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 990.308252] env[61243]: value = "task-1339292" [ 990.308252] env[61243]: _type = "Task" [ 990.308252] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.319996] env[61243]: DEBUG oslo_vmware.api [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339292, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.397921] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339291, 'name': Rename_Task, 'duration_secs': 0.172867} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.398269] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 990.398527] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7356e449-0081-41d3-8622-382a194954d6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.408018] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 990.408018] env[61243]: value = "task-1339293" [ 990.408018] env[61243]: _type = "Task" [ 990.408018] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.445273] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339293, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.543116] env[61243]: DEBUG nova.compute.manager [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 990.569758] env[61243]: DEBUG nova.virt.hardware [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 990.570017] env[61243]: DEBUG nova.virt.hardware [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 990.570193] env[61243]: DEBUG nova.virt.hardware [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 990.570387] env[61243]: DEBUG nova.virt.hardware [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 990.571269] env[61243]: DEBUG nova.virt.hardware [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 990.571448] env[61243]: DEBUG nova.virt.hardware [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 990.571671] env[61243]: DEBUG nova.virt.hardware [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 990.571840] env[61243]: DEBUG nova.virt.hardware [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 990.572014] env[61243]: DEBUG nova.virt.hardware [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 990.572195] env[61243]: DEBUG nova.virt.hardware [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 990.572585] env[61243]: DEBUG nova.virt.hardware [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 990.573461] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805a4b23-a9bb-4dde-9dbe-0fd5e4eb724e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.582026] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc950846-0f32-49cd-a282-ec766956a099 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.738857] env[61243]: DEBUG nova.network.neutron [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 990.803917] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.285s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.804186] env[61243]: DEBUG nova.compute.manager [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=61243) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 990.806933] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.734s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.807175] env[61243]: DEBUG nova.objects.instance [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lazy-loading 'resources' on Instance uuid 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.818601] env[61243]: DEBUG oslo_vmware.api [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339292, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.882049] env[61243]: DEBUG nova.network.neutron [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Updating instance_info_cache with network_info: [{"id": "75bf3cd4-a26f-4ca5-ba11-69971cc910c8", "address": "fa:16:3e:9a:e5:91", "network": {"id": "bce1aecb-e043-4f68-99e2-c6b03c4bb866", "bridge": "br-int", "label": "tempest-ServersTestJSON-652904947-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e641e72e4d2048218ef403456dbb2b82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75bf3cd4-a2", "ovs_interfaceid": "75bf3cd4-a26f-4ca5-ba11-69971cc910c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.916637] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339293, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.931012] env[61243]: DEBUG nova.network.neutron [-] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.953738] env[61243]: DEBUG nova.compute.manager [req-d43a9e3b-d73c-433c-9b71-ab395c5a08bc req-90a17328-424f-4b0e-bf2a-ec4c5c463148 service nova] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Received event network-changed-b2f45b07-0ee1-45ba-a96b-9e8d62f5159d {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.953962] env[61243]: DEBUG nova.compute.manager [req-d43a9e3b-d73c-433c-9b71-ab395c5a08bc req-90a17328-424f-4b0e-bf2a-ec4c5c463148 service nova] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Refreshing instance network info cache due to event network-changed-b2f45b07-0ee1-45ba-a96b-9e8d62f5159d. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 990.954218] env[61243]: DEBUG oslo_concurrency.lockutils [req-d43a9e3b-d73c-433c-9b71-ab395c5a08bc req-90a17328-424f-4b0e-bf2a-ec4c5c463148 service nova] Acquiring lock "refresh_cache-55f24903-b0b4-4124-afde-f31b2680a53a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.954701] env[61243]: DEBUG oslo_concurrency.lockutils [req-d43a9e3b-d73c-433c-9b71-ab395c5a08bc req-90a17328-424f-4b0e-bf2a-ec4c5c463148 service nova] Acquired lock "refresh_cache-55f24903-b0b4-4124-afde-f31b2680a53a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.954701] env[61243]: DEBUG nova.network.neutron [req-d43a9e3b-d73c-433c-9b71-ab395c5a08bc req-90a17328-424f-4b0e-bf2a-ec4c5c463148 service nova] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Refreshing network info cache for port b2f45b07-0ee1-45ba-a96b-9e8d62f5159d {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.168188] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "interface-b0b5b215-d7ff-4960-9d39-cf792d8d4656-317b0929-bd21-499a-aeb7-39bb4b38b703" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.168518] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-b0b5b215-d7ff-4960-9d39-cf792d8d4656-317b0929-bd21-499a-aeb7-39bb4b38b703" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.168967] env[61243]: DEBUG nova.objects.instance [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lazy-loading 'flavor' on Instance uuid b0b5b215-d7ff-4960-9d39-cf792d8d4656 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.325999] env[61243]: DEBUG oslo_vmware.api [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339292, 'name': PowerOffVM_Task, 'duration_secs': 0.902457} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.327296] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 991.327497] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 991.329219] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e98fd417-1511-4c56-8a21-6f363c29c286 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.372143] env[61243]: INFO nova.scheduler.client.report [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Deleted allocation for migration 172672ef-2a6b-494c-85d6-065c551cbfca [ 991.384524] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "refresh_cache-818476f0-e928-44d5-b3d4-101b7a05c9aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.385719] env[61243]: DEBUG nova.compute.manager [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Instance network_info: |[{"id": "75bf3cd4-a26f-4ca5-ba11-69971cc910c8", "address": "fa:16:3e:9a:e5:91", "network": {"id": "bce1aecb-e043-4f68-99e2-c6b03c4bb866", "bridge": "br-int", "label": "tempest-ServersTestJSON-652904947-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e641e72e4d2048218ef403456dbb2b82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75bf3cd4-a2", "ovs_interfaceid": "75bf3cd4-a26f-4ca5-ba11-69971cc910c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 991.385719] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9a:e5:91', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '90878b7b-ddb7-4f47-892b-d6e06f73475f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '75bf3cd4-a26f-4ca5-ba11-69971cc910c8', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 991.393697] env[61243]: DEBUG oslo.service.loopingcall [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.396534] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 991.399190] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e72d748e-fd35-4abd-a6d4-3447ab038473 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.425783] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339293, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.427092] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 991.427092] env[61243]: value = "task-1339295" [ 991.427092] env[61243]: _type = "Task" [ 991.427092] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.437597] env[61243]: INFO nova.compute.manager [-] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Took 1.44 seconds to deallocate network for instance. [ 991.437838] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339295, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.489541] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 991.489838] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 991.490011] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Deleting the datastore file [datastore2] 86179fb0-99df-4b10-a815-c19168e9521e {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 991.490255] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6c5251a6-994f-494b-bb59-4abc0a755487 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.496658] env[61243]: DEBUG oslo_vmware.api [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for the task: (returnval){ [ 991.496658] env[61243]: value = "task-1339296" [ 991.496658] env[61243]: _type = "Task" [ 991.496658] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.508811] env[61243]: DEBUG oslo_vmware.api [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339296, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.527712] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-546ee08e-32b3-4a04-9a4a-dd150ed314fd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.535816] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dff411e-e888-4ed4-990d-0ef4d600d4bc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.568697] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cd469bd-bd50-4a08-86dd-6d79d8862eb5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.576345] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76f1105-d48d-4e72-b3fc-1e5863f073af {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.589742] env[61243]: DEBUG nova.compute.provider_tree [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.833039] env[61243]: DEBUG nova.objects.instance [None req-f8a0c39e-a764-4b2e-ac17-6e74ec3375d5 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lazy-loading 'flavor' on Instance uuid 316572e0-c007-42cb-aaf0-3a8cfcaf24aa {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.851785] env[61243]: DEBUG nova.network.neutron [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Successfully updated port: 6fddda84-0767-4709-ac83-fc4e6060da3c {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 991.877751] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f3e90ad5-567f-4d73-931d-0a1d398ab6cd tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 14.057s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.884363] env[61243]: DEBUG nova.objects.instance [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lazy-loading 'pci_requests' on Instance uuid b0b5b215-d7ff-4960-9d39-cf792d8d4656 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.897943] env[61243]: DEBUG nova.network.neutron [req-d43a9e3b-d73c-433c-9b71-ab395c5a08bc req-90a17328-424f-4b0e-bf2a-ec4c5c463148 service nova] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Updated VIF entry in instance network info cache for port b2f45b07-0ee1-45ba-a96b-9e8d62f5159d. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 991.898409] env[61243]: DEBUG nova.network.neutron [req-d43a9e3b-d73c-433c-9b71-ab395c5a08bc req-90a17328-424f-4b0e-bf2a-ec4c5c463148 service nova] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Updating instance_info_cache with network_info: [{"id": "b2f45b07-0ee1-45ba-a96b-9e8d62f5159d", "address": "fa:16:3e:e8:a7:8b", "network": {"id": "c6d4943d-4a8d-41b4-8023-c4f1549d312d", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1045259606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8549626a07e941e1ab03452355a0f116", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2f45b07-0e", "ovs_interfaceid": "b2f45b07-0ee1-45ba-a96b-9e8d62f5159d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.925654] env[61243]: DEBUG oslo_vmware.api [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339293, 'name': PowerOnVM_Task, 'duration_secs': 1.082582} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.925933] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 991.926156] env[61243]: INFO nova.compute.manager [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Took 9.38 seconds to spawn the instance on the hypervisor. [ 991.926344] env[61243]: DEBUG nova.compute.manager [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 991.927122] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904de338-3a41-4489-a362-20ee3a226334 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.941307] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339295, 'name': CreateVM_Task, 'duration_secs': 0.370752} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.941307] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 991.941860] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.942037] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.942356] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 991.942628] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cc7bbb0-4700-4118-ad63-b84984f4daad {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.946892] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.948095] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 991.948095] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529482ed-56e1-70f4-4ac6-e2e01f959e00" [ 991.948095] env[61243]: _type = "Task" [ 991.948095] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.957599] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529482ed-56e1-70f4-4ac6-e2e01f959e00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.005894] env[61243]: DEBUG oslo_vmware.api [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Task: {'id': task-1339296, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.256328} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.006155] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.006346] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 992.006526] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 992.006703] env[61243]: INFO nova.compute.manager [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Took 1.73 seconds to destroy the instance on the hypervisor. [ 992.006940] env[61243]: DEBUG oslo.service.loopingcall [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.007143] env[61243]: DEBUG nova.compute.manager [-] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 992.007239] env[61243]: DEBUG nova.network.neutron [-] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 992.095933] env[61243]: DEBUG nova.scheduler.client.report [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.337677] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a0c39e-a764-4b2e-ac17-6e74ec3375d5 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.338036] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a0c39e-a764-4b2e-ac17-6e74ec3375d5 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.338081] env[61243]: DEBUG nova.network.neutron [None req-f8a0c39e-a764-4b2e-ac17-6e74ec3375d5 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 992.338240] env[61243]: DEBUG nova.objects.instance [None req-f8a0c39e-a764-4b2e-ac17-6e74ec3375d5 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lazy-loading 'info_cache' on Instance uuid 316572e0-c007-42cb-aaf0-3a8cfcaf24aa {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.353773] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "refresh_cache-5f1c246d-2732-4da4-a00b-b92e2990aa67" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.353935] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "refresh_cache-5f1c246d-2732-4da4-a00b-b92e2990aa67" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.354103] env[61243]: DEBUG nova.network.neutron [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 992.386057] env[61243]: DEBUG nova.objects.base [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 992.386285] env[61243]: DEBUG nova.network.neutron [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 992.400903] env[61243]: DEBUG oslo_concurrency.lockutils [req-d43a9e3b-d73c-433c-9b71-ab395c5a08bc req-90a17328-424f-4b0e-bf2a-ec4c5c463148 service nova] Releasing lock "refresh_cache-55f24903-b0b4-4124-afde-f31b2680a53a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.401196] env[61243]: DEBUG nova.compute.manager [req-d43a9e3b-d73c-433c-9b71-ab395c5a08bc req-90a17328-424f-4b0e-bf2a-ec4c5c463148 service nova] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Received event network-vif-deleted-ca1babd2-ee22-456c-9b5f-205155fa93e4 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.401384] env[61243]: INFO nova.compute.manager [req-d43a9e3b-d73c-433c-9b71-ab395c5a08bc req-90a17328-424f-4b0e-bf2a-ec4c5c463148 service nova] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Neutron deleted interface ca1babd2-ee22-456c-9b5f-205155fa93e4; detaching it from the instance and deleting it from the info cache [ 992.401561] env[61243]: DEBUG nova.network.neutron [req-d43a9e3b-d73c-433c-9b71-ab395c5a08bc req-90a17328-424f-4b0e-bf2a-ec4c5c463148 service nova] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.443024] env[61243]: INFO nova.compute.manager [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Took 20.44 seconds to build instance. [ 992.450352] env[61243]: DEBUG nova.policy [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b32f4788825f4a8e82e028cc10347746', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '683c027e4adc4cc4a6b8747a8c8c0de9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 992.463973] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529482ed-56e1-70f4-4ac6-e2e01f959e00, 'name': SearchDatastore_Task, 'duration_secs': 0.02707} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.464307] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.464557] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 992.464801] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.464958] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.465165] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 992.465429] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0cad71cf-4e0e-460f-9ffe-429bc851556a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.474715] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 992.474715] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 992.475832] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3adf81ad-f280-49ea-a5d5-45e984b92d5b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.480877] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 992.480877] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52149ff4-de5d-34b7-3e67-e26500f38e3d" [ 992.480877] env[61243]: _type = "Task" [ 992.480877] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.488404] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52149ff4-de5d-34b7-3e67-e26500f38e3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.601366] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.794s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.603660] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.591s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.610359] env[61243]: INFO nova.compute.claims [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 992.628357] env[61243]: INFO nova.scheduler.client.report [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Deleted allocations for instance 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba [ 992.758779] env[61243]: DEBUG nova.network.neutron [-] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.841341] env[61243]: DEBUG nova.objects.base [None req-f8a0c39e-a764-4b2e-ac17-6e74ec3375d5 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Object Instance<316572e0-c007-42cb-aaf0-3a8cfcaf24aa> lazy-loaded attributes: flavor,info_cache {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 992.889415] env[61243]: DEBUG nova.network.neutron [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 992.906217] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d0d44a26-f6fc-4ba1-9a2f-e82a2d220f34 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.916238] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c40049-23cf-4b48-b25b-275aa1f7139d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.945834] env[61243]: DEBUG oslo_concurrency.lockutils [None req-281ba41f-e713-4ea4-875f-def3b6c85a25 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "5660de64-7152-4832-960e-2deb1c0bc37e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.949s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.946224] env[61243]: DEBUG nova.compute.manager [req-d43a9e3b-d73c-433c-9b71-ab395c5a08bc req-90a17328-424f-4b0e-bf2a-ec4c5c463148 service nova] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Detach interface failed, port_id=ca1babd2-ee22-456c-9b5f-205155fa93e4, reason: Instance 5103569e-a281-4e89-af69-3a070ab17564 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 992.991532] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52149ff4-de5d-34b7-3e67-e26500f38e3d, 'name': SearchDatastore_Task, 'duration_secs': 0.036264} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.992345] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19c8b9c6-69c3-4392-9292-505cbf3b7792 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.997597] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 992.997597] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52564b5d-eafb-fa12-c112-250eb1aa4059" [ 992.997597] env[61243]: _type = "Task" [ 992.997597] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.004930] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52564b5d-eafb-fa12-c112-250eb1aa4059, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.022771] env[61243]: DEBUG nova.network.neutron [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Updating instance_info_cache with network_info: [{"id": "6fddda84-0767-4709-ac83-fc4e6060da3c", "address": "fa:16:3e:b4:9e:11", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fddda84-07", "ovs_interfaceid": "6fddda84-0767-4709-ac83-fc4e6060da3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.089511] env[61243]: DEBUG nova.compute.manager [req-ac26f735-4aa4-406c-8aec-7c22338d2316 req-aac18805-7187-4c31-9c31-8dadec130156 service nova] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Received event network-vif-plugged-6fddda84-0767-4709-ac83-fc4e6060da3c {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.089722] env[61243]: DEBUG oslo_concurrency.lockutils [req-ac26f735-4aa4-406c-8aec-7c22338d2316 req-aac18805-7187-4c31-9c31-8dadec130156 service nova] Acquiring lock "5f1c246d-2732-4da4-a00b-b92e2990aa67-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.089972] env[61243]: DEBUG oslo_concurrency.lockutils [req-ac26f735-4aa4-406c-8aec-7c22338d2316 req-aac18805-7187-4c31-9c31-8dadec130156 service nova] Lock "5f1c246d-2732-4da4-a00b-b92e2990aa67-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.090218] env[61243]: DEBUG oslo_concurrency.lockutils [req-ac26f735-4aa4-406c-8aec-7c22338d2316 req-aac18805-7187-4c31-9c31-8dadec130156 service nova] Lock "5f1c246d-2732-4da4-a00b-b92e2990aa67-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.090475] env[61243]: DEBUG nova.compute.manager [req-ac26f735-4aa4-406c-8aec-7c22338d2316 req-aac18805-7187-4c31-9c31-8dadec130156 service nova] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] No waiting events found dispatching network-vif-plugged-6fddda84-0767-4709-ac83-fc4e6060da3c {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 993.090685] env[61243]: WARNING nova.compute.manager [req-ac26f735-4aa4-406c-8aec-7c22338d2316 req-aac18805-7187-4c31-9c31-8dadec130156 service nova] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Received unexpected event network-vif-plugged-6fddda84-0767-4709-ac83-fc4e6060da3c for instance with vm_state building and task_state spawning. [ 993.090864] env[61243]: DEBUG nova.compute.manager [req-ac26f735-4aa4-406c-8aec-7c22338d2316 req-aac18805-7187-4c31-9c31-8dadec130156 service nova] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Received event network-changed-6fddda84-0767-4709-ac83-fc4e6060da3c {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.091041] env[61243]: DEBUG nova.compute.manager [req-ac26f735-4aa4-406c-8aec-7c22338d2316 req-aac18805-7187-4c31-9c31-8dadec130156 service nova] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Refreshing instance network info cache due to event network-changed-6fddda84-0767-4709-ac83-fc4e6060da3c. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 993.091219] env[61243]: DEBUG oslo_concurrency.lockutils [req-ac26f735-4aa4-406c-8aec-7c22338d2316 req-aac18805-7187-4c31-9c31-8dadec130156 service nova] Acquiring lock "refresh_cache-5f1c246d-2732-4da4-a00b-b92e2990aa67" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.137152] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1b6d5ab1-be2b-4290-98a8-a68c4c4658d2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "99ce6ca3-a478-4ebe-bf1b-42459aacd9ba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.370s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.263314] env[61243]: INFO nova.compute.manager [-] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Took 1.25 seconds to deallocate network for instance. [ 993.507269] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52564b5d-eafb-fa12-c112-250eb1aa4059, 'name': SearchDatastore_Task, 'duration_secs': 0.036984} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.507608] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.507919] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 818476f0-e928-44d5-b3d4-101b7a05c9aa/818476f0-e928-44d5-b3d4-101b7a05c9aa.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 993.508259] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fed5f813-a1f3-4705-9587-5ee362d0c657 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.514944] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 993.514944] env[61243]: value = "task-1339297" [ 993.514944] env[61243]: _type = "Task" [ 993.514944] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.524446] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339297, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.524905] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "refresh_cache-5f1c246d-2732-4da4-a00b-b92e2990aa67" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.525218] env[61243]: DEBUG nova.compute.manager [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Instance network_info: |[{"id": "6fddda84-0767-4709-ac83-fc4e6060da3c", "address": "fa:16:3e:b4:9e:11", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fddda84-07", "ovs_interfaceid": "6fddda84-0767-4709-ac83-fc4e6060da3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 993.525501] env[61243]: DEBUG oslo_concurrency.lockutils [req-ac26f735-4aa4-406c-8aec-7c22338d2316 req-aac18805-7187-4c31-9c31-8dadec130156 service nova] Acquired lock "refresh_cache-5f1c246d-2732-4da4-a00b-b92e2990aa67" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.525682] env[61243]: DEBUG nova.network.neutron [req-ac26f735-4aa4-406c-8aec-7c22338d2316 req-aac18805-7187-4c31-9c31-8dadec130156 service nova] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Refreshing network info cache for port 6fddda84-0767-4709-ac83-fc4e6060da3c {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 993.527031] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:9e:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ff3ecd2f-0b10-4faf-a512-fd7a20c28df1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6fddda84-0767-4709-ac83-fc4e6060da3c', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 993.534292] env[61243]: DEBUG oslo.service.loopingcall [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.536903] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 993.537677] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d556d907-2474-4985-ada3-02fef47c5a67 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.558175] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 993.558175] env[61243]: value = "task-1339298" [ 993.558175] env[61243]: _type = "Task" [ 993.558175] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.566224] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339298, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.586766] env[61243]: DEBUG nova.network.neutron [None req-f8a0c39e-a764-4b2e-ac17-6e74ec3375d5 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating instance_info_cache with network_info: [{"id": "e21d7ddc-a0d3-4026-a937-f392d5e20fa4", "address": "fa:16:3e:56:77:58", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape21d7ddc-a0", "ovs_interfaceid": "e21d7ddc-a0d3-4026-a937-f392d5e20fa4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.770694] env[61243]: DEBUG oslo_concurrency.lockutils [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.797053] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcae0382-591e-4411-ad31-30ab01f08cff {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.805418] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb842c4-c22d-4183-9979-56ce9cb7503b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.840522] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7abbf8bf-10a7-4c2c-8743-afbba2e2930a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.850720] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41abde64-6272-4da4-84d2-44f0f958916b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.866506] env[61243]: DEBUG nova.compute.provider_tree [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.030193] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339297, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467654} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.032583] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 818476f0-e928-44d5-b3d4-101b7a05c9aa/818476f0-e928-44d5-b3d4-101b7a05c9aa.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 994.035289] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 994.035289] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ece49c08-5344-477f-8157-9555816a215e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.053258] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 994.053258] env[61243]: value = "task-1339299" [ 994.053258] env[61243]: _type = "Task" [ 994.053258] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.069724] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339299, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.075827] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339298, 'name': CreateVM_Task, 'duration_secs': 0.457716} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.076108] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 994.077217] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.077494] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.077961] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 994.078373] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28b12e1a-1769-49a4-94ec-7352524e4646 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.085268] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 994.085268] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52259ede-fc4b-9576-ada5-de08466ab122" [ 994.085268] env[61243]: _type = "Task" [ 994.085268] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.090602] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a0c39e-a764-4b2e-ac17-6e74ec3375d5 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "refresh_cache-316572e0-c007-42cb-aaf0-3a8cfcaf24aa" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.101241] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52259ede-fc4b-9576-ada5-de08466ab122, 'name': SearchDatastore_Task, 'duration_secs': 0.010709} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.105455] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.105879] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 994.106220] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.106642] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.107071] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 994.108318] env[61243]: DEBUG nova.network.neutron [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Successfully updated port: 317b0929-bd21-499a-aeb7-39bb4b38b703 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 994.110034] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c592dd71-69ae-42f8-87a6-af4f472c962a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.127058] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 994.127443] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 994.129530] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-453d0926-ec5d-457a-b671-38bdcf9d50bf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.138565] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 994.138565] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a1cd67-660c-db8c-43a4-d0c0c08fcb5b" [ 994.138565] env[61243]: _type = "Task" [ 994.138565] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.151080] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a1cd67-660c-db8c-43a4-d0c0c08fcb5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.307184] env[61243]: DEBUG nova.network.neutron [req-ac26f735-4aa4-406c-8aec-7c22338d2316 req-aac18805-7187-4c31-9c31-8dadec130156 service nova] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Updated VIF entry in instance network info cache for port 6fddda84-0767-4709-ac83-fc4e6060da3c. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 994.307589] env[61243]: DEBUG nova.network.neutron [req-ac26f735-4aa4-406c-8aec-7c22338d2316 req-aac18805-7187-4c31-9c31-8dadec130156 service nova] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Updating instance_info_cache with network_info: [{"id": "6fddda84-0767-4709-ac83-fc4e6060da3c", "address": "fa:16:3e:b4:9e:11", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fddda84-07", "ovs_interfaceid": "6fddda84-0767-4709-ac83-fc4e6060da3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.370233] env[61243]: DEBUG nova.scheduler.client.report [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 994.563548] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339299, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.106997} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.564026] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 994.564696] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e1aeaf-81d1-4077-8d85-44431ef1765c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.587432] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 818476f0-e928-44d5-b3d4-101b7a05c9aa/818476f0-e928-44d5-b3d4-101b7a05c9aa.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.587801] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad4aac07-dab8-4149-8631-8fca707e4a61 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.603191] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a0c39e-a764-4b2e-ac17-6e74ec3375d5 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 994.603427] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4e445b76-da47-493f-86ab-95705ab11de1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.610566] env[61243]: DEBUG oslo_vmware.api [None req-f8a0c39e-a764-4b2e-ac17-6e74ec3375d5 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 994.610566] env[61243]: value = "task-1339300" [ 994.610566] env[61243]: _type = "Task" [ 994.610566] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.611981] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 994.611981] env[61243]: value = "task-1339301" [ 994.611981] env[61243]: _type = "Task" [ 994.611981] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.617025] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.617025] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.617025] env[61243]: DEBUG nova.network.neutron [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 994.627623] env[61243]: DEBUG oslo_vmware.api [None req-f8a0c39e-a764-4b2e-ac17-6e74ec3375d5 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339300, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.634200] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339301, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.652341] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a1cd67-660c-db8c-43a4-d0c0c08fcb5b, 'name': SearchDatastore_Task, 'duration_secs': 0.051304} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.653235] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f04714af-9ec9-48ac-aa00-e83636bdc9bc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.659280] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 994.659280] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52663bf1-2978-d721-d8be-de779644dbfb" [ 994.659280] env[61243]: _type = "Task" [ 994.659280] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.667841] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52663bf1-2978-d721-d8be-de779644dbfb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.803640] env[61243]: DEBUG nova.compute.manager [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Stashing vm_state: active {{(pid=61243) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 994.810219] env[61243]: DEBUG oslo_concurrency.lockutils [req-ac26f735-4aa4-406c-8aec-7c22338d2316 req-aac18805-7187-4c31-9c31-8dadec130156 service nova] Releasing lock "refresh_cache-5f1c246d-2732-4da4-a00b-b92e2990aa67" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.810474] env[61243]: DEBUG nova.compute.manager [req-ac26f735-4aa4-406c-8aec-7c22338d2316 req-aac18805-7187-4c31-9c31-8dadec130156 service nova] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Received event network-vif-deleted-f2248ec9-0230-4aa8-a016-a30ddf4f45f4 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 994.876385] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.273s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.876923] env[61243]: DEBUG nova.compute.manager [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 994.879593] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.933s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.880259] env[61243]: DEBUG nova.objects.instance [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lazy-loading 'resources' on Instance uuid 5103569e-a281-4e89-af69-3a070ab17564 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.126758] env[61243]: DEBUG oslo_vmware.api [None req-f8a0c39e-a764-4b2e-ac17-6e74ec3375d5 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339300, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.131040] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339301, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.169363] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52663bf1-2978-d721-d8be-de779644dbfb, 'name': SearchDatastore_Task, 'duration_secs': 0.015461} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.169652] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.171182] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 5f1c246d-2732-4da4-a00b-b92e2990aa67/5f1c246d-2732-4da4-a00b-b92e2990aa67.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 995.171182] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac8a16b7-3060-433e-afa1-fdf1b7e32487 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.178563] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 995.178563] env[61243]: value = "task-1339302" [ 995.178563] env[61243]: _type = "Task" [ 995.178563] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.182963] env[61243]: WARNING nova.network.neutron [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] 7724b87f-3721-42a5-81ce-27ae98fc99c2 already exists in list: networks containing: ['7724b87f-3721-42a5-81ce-27ae98fc99c2']. ignoring it [ 995.183204] env[61243]: WARNING nova.network.neutron [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] 7724b87f-3721-42a5-81ce-27ae98fc99c2 already exists in list: networks containing: ['7724b87f-3721-42a5-81ce-27ae98fc99c2']. ignoring it [ 995.189973] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339302, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.327380] env[61243]: DEBUG oslo_concurrency.lockutils [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.385913] env[61243]: DEBUG nova.compute.utils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 995.387768] env[61243]: DEBUG nova.compute.manager [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 995.387996] env[61243]: DEBUG nova.network.neutron [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 995.449523] env[61243]: DEBUG nova.policy [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a9e6dc85a31d4ec38efe0c750c49ead1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '113713170fb94b8a8d7985b9a02faf77', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 995.511535] env[61243]: DEBUG nova.compute.manager [req-5775a178-e797-483d-acab-7524cedac2b8 req-8a991857-5b59-496c-a094-fba027181512 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Received event network-vif-plugged-317b0929-bd21-499a-aeb7-39bb4b38b703 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.511778] env[61243]: DEBUG oslo_concurrency.lockutils [req-5775a178-e797-483d-acab-7524cedac2b8 req-8a991857-5b59-496c-a094-fba027181512 service nova] Acquiring lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.512025] env[61243]: DEBUG oslo_concurrency.lockutils [req-5775a178-e797-483d-acab-7524cedac2b8 req-8a991857-5b59-496c-a094-fba027181512 service nova] Lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.512217] env[61243]: DEBUG oslo_concurrency.lockutils [req-5775a178-e797-483d-acab-7524cedac2b8 req-8a991857-5b59-496c-a094-fba027181512 service nova] Lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.512400] env[61243]: DEBUG nova.compute.manager [req-5775a178-e797-483d-acab-7524cedac2b8 req-8a991857-5b59-496c-a094-fba027181512 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] No waiting events found dispatching network-vif-plugged-317b0929-bd21-499a-aeb7-39bb4b38b703 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 995.512570] env[61243]: WARNING nova.compute.manager [req-5775a178-e797-483d-acab-7524cedac2b8 req-8a991857-5b59-496c-a094-fba027181512 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Received unexpected event network-vif-plugged-317b0929-bd21-499a-aeb7-39bb4b38b703 for instance with vm_state active and task_state None. [ 995.512782] env[61243]: DEBUG nova.compute.manager [req-5775a178-e797-483d-acab-7524cedac2b8 req-8a991857-5b59-496c-a094-fba027181512 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Received event network-changed-317b0929-bd21-499a-aeb7-39bb4b38b703 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.512956] env[61243]: DEBUG nova.compute.manager [req-5775a178-e797-483d-acab-7524cedac2b8 req-8a991857-5b59-496c-a094-fba027181512 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Refreshing instance network info cache due to event network-changed-317b0929-bd21-499a-aeb7-39bb4b38b703. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 995.513354] env[61243]: DEBUG oslo_concurrency.lockutils [req-5775a178-e797-483d-acab-7524cedac2b8 req-8a991857-5b59-496c-a094-fba027181512 service nova] Acquiring lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.634122] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521b17c4-4851-47de-8a42-2807807aaedf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.637573] env[61243]: DEBUG oslo_vmware.api [None req-f8a0c39e-a764-4b2e-ac17-6e74ec3375d5 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339300, 'name': PowerOnVM_Task, 'duration_secs': 0.696817} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.637854] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339301, 'name': ReconfigVM_Task, 'duration_secs': 0.669342} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.638183] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a0c39e-a764-4b2e-ac17-6e74ec3375d5 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 995.638409] env[61243]: DEBUG nova.compute.manager [None req-f8a0c39e-a764-4b2e-ac17-6e74ec3375d5 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 995.638722] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 818476f0-e928-44d5-b3d4-101b7a05c9aa/818476f0-e928-44d5-b3d4-101b7a05c9aa.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 995.640978] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6f00da9-40d6-415a-b079-f62c0d589b18 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.644978] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-04f770c7-1d70-442b-8934-4820f8c73328 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.652021] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c01baf2-0677-4b58-b12b-02eeddd959e6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.660349] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 995.660349] env[61243]: value = "task-1339303" [ 995.660349] env[61243]: _type = "Task" [ 995.660349] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.698833] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9547de2-1fbb-4be5-877b-68192026f7cc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.708481] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339303, 'name': Rename_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.711628] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339302, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.716133] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6d0ee7-9be9-4316-908f-badb178572ab {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.735383] env[61243]: DEBUG nova.compute.provider_tree [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.771281] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "dca3301d-ef3e-48ef-920c-866b4a086ea6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.771673] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "dca3301d-ef3e-48ef-920c-866b4a086ea6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.894021] env[61243]: DEBUG nova.compute.manager [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 995.956102] env[61243]: DEBUG nova.network.neutron [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Updating instance_info_cache with network_info: [{"id": "2333ab37-7807-45d2-a74e-05404217df6d", "address": "fa:16:3e:ea:93:91", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2333ab37-78", "ovs_interfaceid": "2333ab37-7807-45d2-a74e-05404217df6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b91562c3-ce88-4ab3-8413-a5bfeb82aa48", "address": "fa:16:3e:2b:fe:34", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb91562c3-ce", "ovs_interfaceid": "b91562c3-ce88-4ab3-8413-a5bfeb82aa48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "317b0929-bd21-499a-aeb7-39bb4b38b703", "address": "fa:16:3e:16:27:45", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap317b0929-bd", "ovs_interfaceid": "317b0929-bd21-499a-aeb7-39bb4b38b703", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.979932] env[61243]: DEBUG nova.network.neutron [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Successfully created port: 611257cf-50e1-491a-beae-8c942398013d {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 996.175927] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339303, 'name': Rename_Task, 'duration_secs': 0.278165} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.176263] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 996.176671] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bbe58aed-9d22-42cc-b4d4-b619ffca7d2c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.184919] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 996.184919] env[61243]: value = "task-1339304" [ 996.184919] env[61243]: _type = "Task" [ 996.184919] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.192761] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339304, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.202127] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339302, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.844335} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.202391] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 5f1c246d-2732-4da4-a00b-b92e2990aa67/5f1c246d-2732-4da4-a00b-b92e2990aa67.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 996.202609] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 996.202867] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-08225ee9-ba26-4fbc-9ee3-c430af1e1602 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.210303] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 996.210303] env[61243]: value = "task-1339305" [ 996.210303] env[61243]: _type = "Task" [ 996.210303] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.217912] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339305, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.238468] env[61243]: DEBUG nova.scheduler.client.report [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.273681] env[61243]: DEBUG nova.compute.manager [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 996.460072] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.460991] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.460991] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.461736] env[61243]: DEBUG oslo_concurrency.lockutils [req-5775a178-e797-483d-acab-7524cedac2b8 req-8a991857-5b59-496c-a094-fba027181512 service nova] Acquired lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.461736] env[61243]: DEBUG nova.network.neutron [req-5775a178-e797-483d-acab-7524cedac2b8 req-8a991857-5b59-496c-a094-fba027181512 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Refreshing network info cache for port 317b0929-bd21-499a-aeb7-39bb4b38b703 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 996.463129] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc79330a-108d-4d83-9d26-c632a617d89f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.480700] env[61243]: DEBUG nova.virt.hardware [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 996.481537] env[61243]: DEBUG nova.virt.hardware [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 996.481537] env[61243]: DEBUG nova.virt.hardware [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 996.481537] env[61243]: DEBUG nova.virt.hardware [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 996.481537] env[61243]: DEBUG nova.virt.hardware [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 996.481709] env[61243]: DEBUG nova.virt.hardware [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 996.482558] env[61243]: DEBUG nova.virt.hardware [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 996.482558] env[61243]: DEBUG nova.virt.hardware [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 996.482558] env[61243]: DEBUG nova.virt.hardware [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 996.482558] env[61243]: DEBUG nova.virt.hardware [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 996.482558] env[61243]: DEBUG nova.virt.hardware [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 996.489242] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Reconfiguring VM to attach interface {{(pid=61243) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 996.494897] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ce922e8-676c-4d5a-adc9-251e8552ed59 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.518872] env[61243]: DEBUG oslo_vmware.api [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 996.518872] env[61243]: value = "task-1339306" [ 996.518872] env[61243]: _type = "Task" [ 996.518872] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.527928] env[61243]: DEBUG oslo_vmware.api [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339306, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.695318] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339304, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.719854] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339305, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.156045} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.720177] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 996.721015] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5babc49d-1bd7-4cd1-b4ea-26f744714af2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.742641] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 5f1c246d-2732-4da4-a00b-b92e2990aa67/5f1c246d-2732-4da4-a00b-b92e2990aa67.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 996.743416] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.864s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.745363] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0a06e45-a53a-462a-ac17-11b5b2469c88 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.761303] env[61243]: DEBUG oslo_concurrency.lockutils [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.991s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.761553] env[61243]: DEBUG nova.objects.instance [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lazy-loading 'resources' on Instance uuid 86179fb0-99df-4b10-a815-c19168e9521e {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.767658] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 996.767658] env[61243]: value = "task-1339307" [ 996.767658] env[61243]: _type = "Task" [ 996.767658] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.780405] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339307, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.783720] env[61243]: INFO nova.scheduler.client.report [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Deleted allocations for instance 5103569e-a281-4e89-af69-3a070ab17564 [ 996.794423] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.905335] env[61243]: DEBUG nova.compute.manager [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 996.934569] env[61243]: DEBUG nova.virt.hardware [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 996.934965] env[61243]: DEBUG nova.virt.hardware [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 996.935280] env[61243]: DEBUG nova.virt.hardware [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 996.935606] env[61243]: DEBUG nova.virt.hardware [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 996.935891] env[61243]: DEBUG nova.virt.hardware [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 996.936109] env[61243]: DEBUG nova.virt.hardware [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 996.936475] env[61243]: DEBUG nova.virt.hardware [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 996.936744] env[61243]: DEBUG nova.virt.hardware [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 996.937036] env[61243]: DEBUG nova.virt.hardware [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 996.937328] env[61243]: DEBUG nova.virt.hardware [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 996.937638] env[61243]: DEBUG nova.virt.hardware [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 996.938990] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc0283b-0321-4f6f-9bf8-a67c2834aa6a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.949047] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ab38c4-e653-4f78-a26e-e1ef755df6e3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.015339] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.015732] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.015982] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.016265] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.016373] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.019011] env[61243]: INFO nova.compute.manager [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Terminating instance [ 997.020962] env[61243]: DEBUG nova.compute.manager [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 997.021224] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 997.022221] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56816db9-d039-487c-89f0-7ee0c4d37dcb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.033320] env[61243]: DEBUG oslo_vmware.api [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339306, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.035396] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 997.035634] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31c8158d-4e1f-4805-a1d7-c746d4715cdc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.051152] env[61243]: DEBUG oslo_vmware.api [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 997.051152] env[61243]: value = "task-1339308" [ 997.051152] env[61243]: _type = "Task" [ 997.051152] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.061852] env[61243]: DEBUG oslo_vmware.api [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339308, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.187944] env[61243]: DEBUG nova.network.neutron [req-5775a178-e797-483d-acab-7524cedac2b8 req-8a991857-5b59-496c-a094-fba027181512 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Updated VIF entry in instance network info cache for port 317b0929-bd21-499a-aeb7-39bb4b38b703. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 997.188497] env[61243]: DEBUG nova.network.neutron [req-5775a178-e797-483d-acab-7524cedac2b8 req-8a991857-5b59-496c-a094-fba027181512 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Updating instance_info_cache with network_info: [{"id": "2333ab37-7807-45d2-a74e-05404217df6d", "address": "fa:16:3e:ea:93:91", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2333ab37-78", "ovs_interfaceid": "2333ab37-7807-45d2-a74e-05404217df6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b91562c3-ce88-4ab3-8413-a5bfeb82aa48", "address": "fa:16:3e:2b:fe:34", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb91562c3-ce", "ovs_interfaceid": "b91562c3-ce88-4ab3-8413-a5bfeb82aa48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "317b0929-bd21-499a-aeb7-39bb4b38b703", "address": "fa:16:3e:16:27:45", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap317b0929-bd", "ovs_interfaceid": "317b0929-bd21-499a-aeb7-39bb4b38b703", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.202129] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339304, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.277971] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339307, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.289998] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9f575fb3-5007-4629-90c7-ea80ea7fab45 tempest-ImagesTestJSON-832020330 tempest-ImagesTestJSON-832020330-project-member] Lock "5103569e-a281-4e89-af69-3a070ab17564" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.954s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.455139] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6042329-f07c-4f6b-b3b5-85d8641638c4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.462451] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d65adac-70e6-4185-b68c-e18cba38e2a1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.497054] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0439b45-d4c2-4f8c-aae2-1ffe75b9494f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.505318] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-284e4eea-7696-4293-af41-69e2d2376382 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.510397] env[61243]: DEBUG nova.network.neutron [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Successfully updated port: 611257cf-50e1-491a-beae-8c942398013d {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 997.523377] env[61243]: DEBUG nova.compute.provider_tree [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.535242] env[61243]: DEBUG oslo_vmware.api [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339306, 'name': ReconfigVM_Task, 'duration_secs': 1.00377} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.535770] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.535995] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Reconfigured VM to attach interface {{(pid=61243) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 997.561111] env[61243]: DEBUG oslo_vmware.api [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339308, 'name': PowerOffVM_Task, 'duration_secs': 0.17581} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.562281] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 997.562483] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 997.563620] env[61243]: DEBUG nova.compute.manager [req-76f1807b-ec47-4cd7-9510-2c55fc540a6c req-c72f2e5f-a5b5-463c-a2f8-783534bb8b71 service nova] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Received event network-vif-plugged-611257cf-50e1-491a-beae-8c942398013d {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.563867] env[61243]: DEBUG oslo_concurrency.lockutils [req-76f1807b-ec47-4cd7-9510-2c55fc540a6c req-c72f2e5f-a5b5-463c-a2f8-783534bb8b71 service nova] Acquiring lock "2c4217cd-af4a-4d64-a422-73955326df03-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.564378] env[61243]: DEBUG oslo_concurrency.lockutils [req-76f1807b-ec47-4cd7-9510-2c55fc540a6c req-c72f2e5f-a5b5-463c-a2f8-783534bb8b71 service nova] Lock "2c4217cd-af4a-4d64-a422-73955326df03-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.564570] env[61243]: DEBUG oslo_concurrency.lockutils [req-76f1807b-ec47-4cd7-9510-2c55fc540a6c req-c72f2e5f-a5b5-463c-a2f8-783534bb8b71 service nova] Lock "2c4217cd-af4a-4d64-a422-73955326df03-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.564750] env[61243]: DEBUG nova.compute.manager [req-76f1807b-ec47-4cd7-9510-2c55fc540a6c req-c72f2e5f-a5b5-463c-a2f8-783534bb8b71 service nova] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] No waiting events found dispatching network-vif-plugged-611257cf-50e1-491a-beae-8c942398013d {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 997.564965] env[61243]: WARNING nova.compute.manager [req-76f1807b-ec47-4cd7-9510-2c55fc540a6c req-c72f2e5f-a5b5-463c-a2f8-783534bb8b71 service nova] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Received unexpected event network-vif-plugged-611257cf-50e1-491a-beae-8c942398013d for instance with vm_state building and task_state spawning. [ 997.565179] env[61243]: DEBUG nova.compute.manager [req-76f1807b-ec47-4cd7-9510-2c55fc540a6c req-c72f2e5f-a5b5-463c-a2f8-783534bb8b71 service nova] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Received event network-changed-611257cf-50e1-491a-beae-8c942398013d {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.565754] env[61243]: DEBUG nova.compute.manager [req-76f1807b-ec47-4cd7-9510-2c55fc540a6c req-c72f2e5f-a5b5-463c-a2f8-783534bb8b71 service nova] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Refreshing instance network info cache due to event network-changed-611257cf-50e1-491a-beae-8c942398013d. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 997.565754] env[61243]: DEBUG oslo_concurrency.lockutils [req-76f1807b-ec47-4cd7-9510-2c55fc540a6c req-c72f2e5f-a5b5-463c-a2f8-783534bb8b71 service nova] Acquiring lock "refresh_cache-2c4217cd-af4a-4d64-a422-73955326df03" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.565754] env[61243]: DEBUG oslo_concurrency.lockutils [req-76f1807b-ec47-4cd7-9510-2c55fc540a6c req-c72f2e5f-a5b5-463c-a2f8-783534bb8b71 service nova] Acquired lock "refresh_cache-2c4217cd-af4a-4d64-a422-73955326df03" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.565897] env[61243]: DEBUG nova.network.neutron [req-76f1807b-ec47-4cd7-9510-2c55fc540a6c req-c72f2e5f-a5b5-463c-a2f8-783534bb8b71 service nova] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Refreshing network info cache for port 611257cf-50e1-491a-beae-8c942398013d {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 997.566807] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-40e18e13-643a-4d8b-b222-b7718ae7d57b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.630828] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 997.631103] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 997.631296] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Deleting the datastore file [datastore1] 316572e0-c007-42cb-aaf0-3a8cfcaf24aa {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 997.631561] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f1f2621b-32a9-4dc7-a2b2-98d0fe05291c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.638277] env[61243]: DEBUG oslo_vmware.api [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 997.638277] env[61243]: value = "task-1339310" [ 997.638277] env[61243]: _type = "Task" [ 997.638277] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.646175] env[61243]: DEBUG oslo_vmware.api [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339310, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.694703] env[61243]: DEBUG oslo_concurrency.lockutils [req-5775a178-e797-483d-acab-7524cedac2b8 req-8a991857-5b59-496c-a094-fba027181512 service nova] Releasing lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.699477] env[61243]: DEBUG oslo_vmware.api [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339304, 'name': PowerOnVM_Task, 'duration_secs': 1.258369} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.699741] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 997.699985] env[61243]: INFO nova.compute.manager [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Took 9.43 seconds to spawn the instance on the hypervisor. [ 997.700219] env[61243]: DEBUG nova.compute.manager [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 997.701034] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-285ca309-8d33-4715-8ff4-4667c1c2437a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.781903] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339307, 'name': ReconfigVM_Task, 'duration_secs': 0.540236} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.782367] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 5f1c246d-2732-4da4-a00b-b92e2990aa67/5f1c246d-2732-4da4-a00b-b92e2990aa67.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.783367] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2be6a691-ab3a-428b-b7d0-8ee5fef5fb00 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.791120] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 997.791120] env[61243]: value = "task-1339311" [ 997.791120] env[61243]: _type = "Task" [ 997.791120] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.802777] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339311, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.013413] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "refresh_cache-2c4217cd-af4a-4d64-a422-73955326df03" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.030921] env[61243]: DEBUG nova.scheduler.client.report [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.040346] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b729881c-73a7-4180-a200-097b21ae1270 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-b0b5b215-d7ff-4960-9d39-cf792d8d4656-317b0929-bd21-499a-aeb7-39bb4b38b703" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.872s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.104108] env[61243]: DEBUG nova.network.neutron [req-76f1807b-ec47-4cd7-9510-2c55fc540a6c req-c72f2e5f-a5b5-463c-a2f8-783534bb8b71 service nova] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 998.150255] env[61243]: DEBUG oslo_vmware.api [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339310, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.347105} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.152412] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 998.152618] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 998.152932] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 998.153204] env[61243]: INFO nova.compute.manager [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Took 1.13 seconds to destroy the instance on the hypervisor. [ 998.153529] env[61243]: DEBUG oslo.service.loopingcall [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 998.153753] env[61243]: DEBUG nova.compute.manager [-] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 998.153852] env[61243]: DEBUG nova.network.neutron [-] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 998.216122] env[61243]: DEBUG nova.network.neutron [req-76f1807b-ec47-4cd7-9510-2c55fc540a6c req-c72f2e5f-a5b5-463c-a2f8-783534bb8b71 service nova] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.218780] env[61243]: INFO nova.compute.manager [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Took 24.33 seconds to build instance. [ 998.936671] env[61243]: DEBUG oslo_concurrency.lockutils [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.175s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.939886] env[61243]: DEBUG oslo_concurrency.lockutils [req-76f1807b-ec47-4cd7-9510-2c55fc540a6c req-c72f2e5f-a5b5-463c-a2f8-783534bb8b71 service nova] Releasing lock "refresh_cache-2c4217cd-af4a-4d64-a422-73955326df03" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.939886] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d8a60492-4229-4ca7-a38e-9409718340a0 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "818476f0-e928-44d5-b3d4-101b7a05c9aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.062s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.945780] env[61243]: DEBUG oslo_concurrency.lockutils [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.619s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.947545] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "refresh_cache-2c4217cd-af4a-4d64-a422-73955326df03" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.947716] env[61243]: DEBUG nova.network.neutron [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 998.958150] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339311, 'name': Rename_Task, 'duration_secs': 0.187855} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.958481] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 998.958759] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7eedc93b-8e2c-4216-b320-5ff7a78d524b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.967287] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 998.967287] env[61243]: value = "task-1339312" [ 998.967287] env[61243]: _type = "Task" [ 998.967287] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.976463] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339312, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.981629] env[61243]: INFO nova.scheduler.client.report [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Deleted allocations for instance 86179fb0-99df-4b10-a815-c19168e9521e [ 999.447019] env[61243]: DEBUG nova.network.neutron [-] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.447019] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "818476f0-e928-44d5-b3d4-101b7a05c9aa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.447019] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "818476f0-e928-44d5-b3d4-101b7a05c9aa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.447019] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "818476f0-e928-44d5-b3d4-101b7a05c9aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.447019] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "818476f0-e928-44d5-b3d4-101b7a05c9aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.447019] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "818476f0-e928-44d5-b3d4-101b7a05c9aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.449323] env[61243]: INFO nova.compute.manager [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Terminating instance [ 999.452661] env[61243]: INFO nova.compute.claims [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 999.459781] env[61243]: DEBUG nova.compute.manager [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 999.459781] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 999.460288] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdce8046-49df-4888-bbe7-9dddcfe95595 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.468945] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 999.472187] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-358d1b84-640f-4878-9acf-d02199682957 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.479481] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339312, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.480771] env[61243]: DEBUG oslo_vmware.api [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 999.480771] env[61243]: value = "task-1339313" [ 999.480771] env[61243]: _type = "Task" [ 999.480771] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.490183] env[61243]: DEBUG oslo_concurrency.lockutils [None req-194624fa-509c-46ff-870e-c7a6adb4c0d3 tempest-ServersNegativeTestJSON-1743523096 tempest-ServersNegativeTestJSON-1743523096-project-member] Lock "86179fb0-99df-4b10-a815-c19168e9521e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.221s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.493980] env[61243]: DEBUG oslo_vmware.api [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339313, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.510623] env[61243]: DEBUG nova.network.neutron [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 999.644775] env[61243]: DEBUG nova.compute.manager [req-14aea7e2-43ae-46a4-90ec-e6ff21820601 req-363c35a8-664b-43b8-a008-b82b8dc4a691 service nova] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Received event network-vif-deleted-e21d7ddc-a0d3-4026-a937-f392d5e20fa4 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 999.949742] env[61243]: INFO nova.compute.manager [-] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Took 1.80 seconds to deallocate network for instance. [ 999.961663] env[61243]: INFO nova.compute.resource_tracker [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Updating resource usage from migration 50690685-915d-4db4-8806-bb464dd29d17 [ 999.980159] env[61243]: DEBUG nova.network.neutron [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Updating instance_info_cache with network_info: [{"id": "611257cf-50e1-491a-beae-8c942398013d", "address": "fa:16:3e:6a:c8:a1", "network": {"id": "0788f9ac-5a15-4431-a711-9b4f6cf79fbc", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1210957657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "113713170fb94b8a8d7985b9a02faf77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c7821ea-f92f-4f06-a4cb-05e1186a9d22", "external-id": "nsx-vlan-transportzone-69", "segmentation_id": 69, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap611257cf-50", "ovs_interfaceid": "611257cf-50e1-491a-beae-8c942398013d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.997056] env[61243]: DEBUG oslo_vmware.api [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339312, 'name': PowerOnVM_Task, 'duration_secs': 0.937187} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.004024] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1000.004024] env[61243]: INFO nova.compute.manager [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Took 9.46 seconds to spawn the instance on the hypervisor. [ 1000.004024] env[61243]: DEBUG nova.compute.manager [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1000.004024] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c00a1c-840f-4440-bbf3-370db1173df2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.013185] env[61243]: DEBUG oslo_vmware.api [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339313, 'name': PowerOffVM_Task, 'duration_secs': 0.46725} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.014189] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1000.014528] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1000.015188] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-92bcc426-5bc5-492f-b609-c4bc067bff49 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.106873] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1000.107328] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1000.107584] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleting the datastore file [datastore1] 818476f0-e928-44d5-b3d4-101b7a05c9aa {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1000.107871] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c769d27c-7f12-4a81-90e6-e025425c72e0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.117463] env[61243]: DEBUG oslo_vmware.api [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1000.117463] env[61243]: value = "task-1339315" [ 1000.117463] env[61243]: _type = "Task" [ 1000.117463] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.127718] env[61243]: DEBUG oslo_vmware.api [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339315, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.175552] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "interface-b0b5b215-d7ff-4960-9d39-cf792d8d4656-b91562c3-ce88-4ab3-8413-a5bfeb82aa48" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.175874] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-b0b5b215-d7ff-4960-9d39-cf792d8d4656-b91562c3-ce88-4ab3-8413-a5bfeb82aa48" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.199676] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec000a05-24a8-41d0-a0db-0102f762f599 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.207679] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0997528-230f-4f83-9dc2-d3438293695f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.237786] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaa940e8-e33a-4e6e-b1cb-0f3e416984c6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.245522] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7437f9a6-7035-4db6-b841-65bda0dd0ce2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.260086] env[61243]: DEBUG nova.compute.provider_tree [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.460574] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.499964] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "refresh_cache-2c4217cd-af4a-4d64-a422-73955326df03" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.499964] env[61243]: DEBUG nova.compute.manager [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Instance network_info: |[{"id": "611257cf-50e1-491a-beae-8c942398013d", "address": "fa:16:3e:6a:c8:a1", "network": {"id": "0788f9ac-5a15-4431-a711-9b4f6cf79fbc", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1210957657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "113713170fb94b8a8d7985b9a02faf77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c7821ea-f92f-4f06-a4cb-05e1186a9d22", "external-id": "nsx-vlan-transportzone-69", "segmentation_id": 69, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap611257cf-50", "ovs_interfaceid": "611257cf-50e1-491a-beae-8c942398013d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1000.499964] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:c8:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5c7821ea-f92f-4f06-a4cb-05e1186a9d22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '611257cf-50e1-491a-beae-8c942398013d', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1000.508470] env[61243]: DEBUG oslo.service.loopingcall [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.508716] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1000.509295] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a82cba37-71af-491e-8822-47a124acc065 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.534474] env[61243]: INFO nova.compute.manager [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Took 24.58 seconds to build instance. [ 1000.540395] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1000.540395] env[61243]: value = "task-1339316" [ 1000.540395] env[61243]: _type = "Task" [ 1000.540395] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.546656] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339316, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.628237] env[61243]: DEBUG oslo_vmware.api [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339315, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145967} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.628643] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.628743] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1000.628928] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1000.629426] env[61243]: INFO nova.compute.manager [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1000.629426] env[61243]: DEBUG oslo.service.loopingcall [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.629566] env[61243]: DEBUG nova.compute.manager [-] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1000.629659] env[61243]: DEBUG nova.network.neutron [-] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1000.679389] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.679572] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.680515] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d091cec-88fb-493b-ac48-ef58698899ce {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.704309] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087bbcd9-7b54-40d6-8042-36efc6922c9a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.739607] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Reconfiguring VM to detach interface {{(pid=61243) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1000.741890] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c3c285c-7d48-4861-b636-33633c781dc2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.760160] env[61243]: DEBUG oslo_vmware.api [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1000.760160] env[61243]: value = "task-1339317" [ 1000.760160] env[61243]: _type = "Task" [ 1000.760160] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.764743] env[61243]: DEBUG nova.scheduler.client.report [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.773171] env[61243]: DEBUG oslo_vmware.api [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339317, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.038680] env[61243]: DEBUG oslo_concurrency.lockutils [None req-5e70a03d-42e5-4d93-982b-2f6f2cc23acd tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "5f1c246d-2732-4da4-a00b-b92e2990aa67" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.092s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.058498] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339316, 'name': CreateVM_Task, 'duration_secs': 0.390747} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.058837] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1001.059623] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.059623] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.060037] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1001.061603] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a386254-4f58-4308-9de2-5df75e30ee4b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.067837] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 1001.067837] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52568355-e63a-3384-fc43-67d409f38012" [ 1001.067837] env[61243]: _type = "Task" [ 1001.067837] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.078374] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52568355-e63a-3384-fc43-67d409f38012, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.272069] env[61243]: DEBUG oslo_concurrency.lockutils [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.324s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.272069] env[61243]: INFO nova.compute.manager [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Migrating [ 1001.286083] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.491s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.287358] env[61243]: INFO nova.compute.claims [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1001.289913] env[61243]: DEBUG oslo_vmware.api [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.583770] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52568355-e63a-3384-fc43-67d409f38012, 'name': SearchDatastore_Task, 'duration_secs': 0.010478} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.583770] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.584790] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1001.584790] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.584790] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.584790] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1001.584790] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5091acb9-2385-46e5-9fd7-c8867a1530c1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.589728] env[61243]: DEBUG nova.network.neutron [-] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.596772] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1001.596966] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1001.597743] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45463911-2311-4600-9718-5e0fa53e11b0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.604145] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 1001.604145] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5246eadf-6813-669f-091a-cd8b4689dfd8" [ 1001.604145] env[61243]: _type = "Task" [ 1001.604145] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.612511] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5246eadf-6813-669f-091a-cd8b4689dfd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.672263] env[61243]: DEBUG nova.compute.manager [req-f2835d62-b83a-4207-b047-e5874b32bc7a req-d1f821f4-9234-4507-b348-6422ebb0eeb0 service nova] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Received event network-vif-deleted-75bf3cd4-a26f-4ca5-ba11-69971cc910c8 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1001.774875] env[61243]: DEBUG oslo_vmware.api [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.799969] env[61243]: DEBUG oslo_concurrency.lockutils [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "refresh_cache-5660de64-7152-4832-960e-2deb1c0bc37e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.800350] env[61243]: DEBUG oslo_concurrency.lockutils [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "refresh_cache-5660de64-7152-4832-960e-2deb1c0bc37e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.800830] env[61243]: DEBUG nova.network.neutron [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1002.092013] env[61243]: INFO nova.compute.manager [-] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Took 1.46 seconds to deallocate network for instance. [ 1002.116732] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5246eadf-6813-669f-091a-cd8b4689dfd8, 'name': SearchDatastore_Task, 'duration_secs': 0.010095} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.117911] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3e50232-a662-4832-a283-5df27e1183d2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.124629] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 1002.124629] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524676cd-d020-8d3c-2ff0-8da03e613068" [ 1002.124629] env[61243]: _type = "Task" [ 1002.124629] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.133135] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524676cd-d020-8d3c-2ff0-8da03e613068, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.272421] env[61243]: DEBUG oslo_vmware.api [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.513650] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e3bea4-cea5-4d69-bd9a-a891fd662b72 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.521783] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e629d3-3624-40df-a700-288b96fac09e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.558753] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7979c2d3-daab-4576-b8a5-dfb005c3dc87 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.567135] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f85c671-3e08-4cb0-8256-c77a94df39a3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.581571] env[61243]: DEBUG nova.compute.provider_tree [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.598803] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.605188] env[61243]: DEBUG nova.network.neutron [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Updating instance_info_cache with network_info: [{"id": "2b065192-815c-492f-8f79-bfa9553ff080", "address": "fa:16:3e:94:3e:9e", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b065192-81", "ovs_interfaceid": "2b065192-815c-492f-8f79-bfa9553ff080", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.638227] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524676cd-d020-8d3c-2ff0-8da03e613068, 'name': SearchDatastore_Task, 'duration_secs': 0.009387} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.638506] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.639357] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 2c4217cd-af4a-4d64-a422-73955326df03/2c4217cd-af4a-4d64-a422-73955326df03.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1002.639357] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-678ff1b7-31d2-49de-9cbb-939c9d6105fd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.645953] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 1002.645953] env[61243]: value = "task-1339318" [ 1002.645953] env[61243]: _type = "Task" [ 1002.645953] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.657319] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339318, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.771427] env[61243]: DEBUG oslo_vmware.api [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.085506] env[61243]: DEBUG nova.scheduler.client.report [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.106433] env[61243]: DEBUG oslo_concurrency.lockutils [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "refresh_cache-5660de64-7152-4832-960e-2deb1c0bc37e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.159333] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339318, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481789} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.160501] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 2c4217cd-af4a-4d64-a422-73955326df03/2c4217cd-af4a-4d64-a422-73955326df03.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1003.160501] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1003.160983] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-82aa3a51-f80a-404e-bfa4-e2d93274a045 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.168681] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 1003.168681] env[61243]: value = "task-1339319" [ 1003.168681] env[61243]: _type = "Task" [ 1003.168681] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.176012] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339319, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.271704] env[61243]: DEBUG oslo_vmware.api [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.591189] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.305s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.591697] env[61243]: DEBUG nova.compute.manager [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1003.594706] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.135s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.595006] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.597163] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.999s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.597543] env[61243]: DEBUG nova.objects.instance [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lazy-loading 'resources' on Instance uuid 818476f0-e928-44d5-b3d4-101b7a05c9aa {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.619727] env[61243]: INFO nova.scheduler.client.report [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Deleted allocations for instance 316572e0-c007-42cb-aaf0-3a8cfcaf24aa [ 1003.677726] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339319, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.353187} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.678821] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1003.679542] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c56d99-67f2-4cf4-a7c4-727596cc0363 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.704220] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 2c4217cd-af4a-4d64-a422-73955326df03/2c4217cd-af4a-4d64-a422-73955326df03.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1003.704220] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aed5ce18-ccac-46f5-89c9-d18e73480468 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.720964] env[61243]: DEBUG nova.compute.manager [req-3891e4f3-b3ab-4d14-a45d-ad8b68b18441 req-7a922f8c-c59b-46fd-98c6-29927791c4ef service nova] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Received event network-changed-6fddda84-0767-4709-ac83-fc4e6060da3c {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1003.721207] env[61243]: DEBUG nova.compute.manager [req-3891e4f3-b3ab-4d14-a45d-ad8b68b18441 req-7a922f8c-c59b-46fd-98c6-29927791c4ef service nova] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Refreshing instance network info cache due to event network-changed-6fddda84-0767-4709-ac83-fc4e6060da3c. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1003.721417] env[61243]: DEBUG oslo_concurrency.lockutils [req-3891e4f3-b3ab-4d14-a45d-ad8b68b18441 req-7a922f8c-c59b-46fd-98c6-29927791c4ef service nova] Acquiring lock "refresh_cache-5f1c246d-2732-4da4-a00b-b92e2990aa67" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.721567] env[61243]: DEBUG oslo_concurrency.lockutils [req-3891e4f3-b3ab-4d14-a45d-ad8b68b18441 req-7a922f8c-c59b-46fd-98c6-29927791c4ef service nova] Acquired lock "refresh_cache-5f1c246d-2732-4da4-a00b-b92e2990aa67" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.721728] env[61243]: DEBUG nova.network.neutron [req-3891e4f3-b3ab-4d14-a45d-ad8b68b18441 req-7a922f8c-c59b-46fd-98c6-29927791c4ef service nova] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Refreshing network info cache for port 6fddda84-0767-4709-ac83-fc4e6060da3c {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1003.724946] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 1003.724946] env[61243]: value = "task-1339320" [ 1003.724946] env[61243]: _type = "Task" [ 1003.724946] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.734119] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339320, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.774460] env[61243]: DEBUG oslo_vmware.api [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.102490] env[61243]: DEBUG nova.compute.utils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1004.108418] env[61243]: DEBUG nova.compute.manager [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1004.108418] env[61243]: DEBUG nova.network.neutron [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1004.128355] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b581f7e-402f-4bcd-85de-553946e64318 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "316572e0-c007-42cb-aaf0-3a8cfcaf24aa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.112s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.174321] env[61243]: DEBUG nova.policy [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32d9e3276f974c8f87573270909076ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7651b565d1647eb948fe07faba32a41', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1004.239831] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339320, 'name': ReconfigVM_Task, 'duration_secs': 0.27137} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.240670] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 2c4217cd-af4a-4d64-a422-73955326df03/2c4217cd-af4a-4d64-a422-73955326df03.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1004.241571] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c4b1568-4682-41d4-a513-50ff8ed73c99 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.252082] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 1004.252082] env[61243]: value = "task-1339321" [ 1004.252082] env[61243]: _type = "Task" [ 1004.252082] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.258637] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339321, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.277035] env[61243]: DEBUG oslo_vmware.api [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.317136] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6df6a7-1d2e-478c-b366-f77c0e2a0fa4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.324737] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6975eb-a08d-4895-9ea6-d2010c2192d4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.359935] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d086f2-f1a5-4939-8570-6f16a34af63c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.368588] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30c1b5d-f75b-4b42-b512-9ba5dd224f06 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.382227] env[61243]: DEBUG nova.compute.provider_tree [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1004.609311] env[61243]: DEBUG nova.compute.manager [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1004.619272] env[61243]: DEBUG nova.network.neutron [req-3891e4f3-b3ab-4d14-a45d-ad8b68b18441 req-7a922f8c-c59b-46fd-98c6-29927791c4ef service nova] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Updated VIF entry in instance network info cache for port 6fddda84-0767-4709-ac83-fc4e6060da3c. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1004.619651] env[61243]: DEBUG nova.network.neutron [req-3891e4f3-b3ab-4d14-a45d-ad8b68b18441 req-7a922f8c-c59b-46fd-98c6-29927791c4ef service nova] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Updating instance_info_cache with network_info: [{"id": "6fddda84-0767-4709-ac83-fc4e6060da3c", "address": "fa:16:3e:b4:9e:11", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fddda84-07", "ovs_interfaceid": "6fddda84-0767-4709-ac83-fc4e6060da3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.625710] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7343c117-8647-4e21-bb46-5ee2b0c6d19a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.644187] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Updating instance '5660de64-7152-4832-960e-2deb1c0bc37e' progress to 0 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1004.651223] env[61243]: DEBUG nova.network.neutron [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Successfully created port: ed08f6b9-7027-417b-91d5-e9639b3fed92 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1004.760771] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339321, 'name': Rename_Task, 'duration_secs': 0.178482} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.761192] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1004.761391] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2b949ebc-4352-4a54-926e-d57a5b2a5683 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.770230] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 1004.770230] env[61243]: value = "task-1339322" [ 1004.770230] env[61243]: _type = "Task" [ 1004.770230] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.778794] env[61243]: DEBUG oslo_vmware.api [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.784256] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339322, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.909657] env[61243]: ERROR nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [req-2cafe0cd-ea28-4b82-9c2c-c70cc543c023] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2cafe0cd-ea28-4b82-9c2c-c70cc543c023"}]} [ 1004.932612] env[61243]: DEBUG nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1004.953936] env[61243]: DEBUG nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1004.954525] env[61243]: DEBUG nova.compute.provider_tree [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1004.976952] env[61243]: DEBUG nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1005.004515] env[61243]: DEBUG nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1005.059769] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Acquiring lock "0e03eee3-e7bc-483f-948f-a5b408375e24" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.061032] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Lock "0e03eee3-e7bc-483f-948f-a5b408375e24" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.123696] env[61243]: DEBUG oslo_concurrency.lockutils [req-3891e4f3-b3ab-4d14-a45d-ad8b68b18441 req-7a922f8c-c59b-46fd-98c6-29927791c4ef service nova] Releasing lock "refresh_cache-5f1c246d-2732-4da4-a00b-b92e2990aa67" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.150751] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1005.152180] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cdb25e48-55eb-4059-992f-087935f06b7c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.160673] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1005.160673] env[61243]: value = "task-1339323" [ 1005.160673] env[61243]: _type = "Task" [ 1005.160673] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.176080] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339323, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.209715] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64868211-3cf7-49e3-95c6-034d6a4d8c5f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.218113] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c1b34f-62b0-460a-b9c3-46cd47ec97fc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.254110] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e47c7a37-47b6-4212-bac9-a3ac543cd487 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.262294] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab94be8-a73c-4fe0-b220-c3e1e603a17c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.275100] env[61243]: DEBUG oslo_vmware.api [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.286067] env[61243]: DEBUG nova.compute.provider_tree [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1005.292597] env[61243]: DEBUG oslo_vmware.api [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339322, 'name': PowerOnVM_Task, 'duration_secs': 0.49781} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.292738] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1005.292878] env[61243]: INFO nova.compute.manager [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Took 8.39 seconds to spawn the instance on the hypervisor. [ 1005.293471] env[61243]: DEBUG nova.compute.manager [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1005.294066] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2195addc-afb7-40d3-a7f0-91f4e823c385 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.324085] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "c1a0d242-4eae-4c03-8341-840b41341f17" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.324386] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "c1a0d242-4eae-4c03-8341-840b41341f17" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.565159] env[61243]: DEBUG nova.compute.manager [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1005.622481] env[61243]: DEBUG nova.compute.manager [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1005.647960] env[61243]: DEBUG nova.virt.hardware [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.648230] env[61243]: DEBUG nova.virt.hardware [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.648398] env[61243]: DEBUG nova.virt.hardware [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.648587] env[61243]: DEBUG nova.virt.hardware [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.648736] env[61243]: DEBUG nova.virt.hardware [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.648889] env[61243]: DEBUG nova.virt.hardware [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.649114] env[61243]: DEBUG nova.virt.hardware [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.649284] env[61243]: DEBUG nova.virt.hardware [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.649458] env[61243]: DEBUG nova.virt.hardware [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.649627] env[61243]: DEBUG nova.virt.hardware [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.649805] env[61243]: DEBUG nova.virt.hardware [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.650675] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29797fb-f40e-41f6-bfbf-a9da2ed3d2ae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.658527] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b98e35ff-574f-44c6-8d63-f5db95a0d01b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.670095] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339323, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.776642] env[61243]: DEBUG oslo_vmware.api [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.807455] env[61243]: ERROR nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [req-cf61a632-d4f4-4df3-b9b7-9c22a3134d9a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-cf61a632-d4f4-4df3-b9b7-9c22a3134d9a"}]} [ 1005.820287] env[61243]: INFO nova.compute.manager [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Took 16.83 seconds to build instance. [ 1005.827310] env[61243]: INFO nova.compute.manager [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Detaching volume bc3b5837-80e8-4d8a-8a68-5ec6c3b89880 [ 1005.840113] env[61243]: DEBUG nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1005.862131] env[61243]: DEBUG nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1005.862409] env[61243]: DEBUG nova.compute.provider_tree [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1005.871775] env[61243]: INFO nova.virt.block_device [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Attempting to driver detach volume bc3b5837-80e8-4d8a-8a68-5ec6c3b89880 from mountpoint /dev/sdb [ 1005.872036] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Volume detach. Driver type: vmdk {{(pid=61243) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1005.872245] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285747', 'volume_id': 'bc3b5837-80e8-4d8a-8a68-5ec6c3b89880', 'name': 'volume-bc3b5837-80e8-4d8a-8a68-5ec6c3b89880', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'c1a0d242-4eae-4c03-8341-840b41341f17', 'attached_at': '', 'detached_at': '', 'volume_id': 'bc3b5837-80e8-4d8a-8a68-5ec6c3b89880', 'serial': 'bc3b5837-80e8-4d8a-8a68-5ec6c3b89880'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1005.873215] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35984b82-32af-499d-a5b7-0f7242228d1e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.879538] env[61243]: DEBUG nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1005.901552] env[61243]: DEBUG nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1005.905604] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f9bd855-3fe2-47a9-81a9-232db0db44e4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.914909] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4431ebee-8614-451e-8d0e-961208c3f01c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.954365] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d26a6a20-a01c-4395-8c00-ca921bd877da {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.964604] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.964939] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.981197] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] The volume has not been displaced from its original location: [datastore2] volume-bc3b5837-80e8-4d8a-8a68-5ec6c3b89880/volume-bc3b5837-80e8-4d8a-8a68-5ec6c3b89880.vmdk. No consolidation needed. {{(pid=61243) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1005.986494] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Reconfiguring VM instance instance-00000046 to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1005.990139] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4005cd66-37e0-4fd7-9967-8a37d5c48deb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.008861] env[61243]: DEBUG oslo_vmware.api [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1006.008861] env[61243]: value = "task-1339324" [ 1006.008861] env[61243]: _type = "Task" [ 1006.008861] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.022877] env[61243]: DEBUG oslo_vmware.api [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339324, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.096941] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.119760] env[61243]: DEBUG nova.compute.manager [req-2b8f9c1e-dcce-4063-a949-ad86fd794bdc req-2152c96d-72e9-48b3-ba55-ac808899a163 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Received event network-vif-plugged-ed08f6b9-7027-417b-91d5-e9639b3fed92 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.119760] env[61243]: DEBUG oslo_concurrency.lockutils [req-2b8f9c1e-dcce-4063-a949-ad86fd794bdc req-2152c96d-72e9-48b3-ba55-ac808899a163 service nova] Acquiring lock "dca3301d-ef3e-48ef-920c-866b4a086ea6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.120103] env[61243]: DEBUG oslo_concurrency.lockutils [req-2b8f9c1e-dcce-4063-a949-ad86fd794bdc req-2152c96d-72e9-48b3-ba55-ac808899a163 service nova] Lock "dca3301d-ef3e-48ef-920c-866b4a086ea6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.120977] env[61243]: DEBUG oslo_concurrency.lockutils [req-2b8f9c1e-dcce-4063-a949-ad86fd794bdc req-2152c96d-72e9-48b3-ba55-ac808899a163 service nova] Lock "dca3301d-ef3e-48ef-920c-866b4a086ea6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.120977] env[61243]: DEBUG nova.compute.manager [req-2b8f9c1e-dcce-4063-a949-ad86fd794bdc req-2152c96d-72e9-48b3-ba55-ac808899a163 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] No waiting events found dispatching network-vif-plugged-ed08f6b9-7027-417b-91d5-e9639b3fed92 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1006.120977] env[61243]: WARNING nova.compute.manager [req-2b8f9c1e-dcce-4063-a949-ad86fd794bdc req-2152c96d-72e9-48b3-ba55-ac808899a163 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Received unexpected event network-vif-plugged-ed08f6b9-7027-417b-91d5-e9639b3fed92 for instance with vm_state building and task_state spawning. [ 1006.175421] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339323, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.183491] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2310a3e5-8e49-4a67-899e-4e4a952f63a3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.190808] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eacc45c7-dbf1-4e84-870e-722f6f7dc7fa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.224386] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c0bfe27-20bf-4277-84c2-d2e0c62a4862 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.232287] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80f14a87-54be-4358-8217-f5a3bb1406fa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.251032] env[61243]: DEBUG nova.compute.provider_tree [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1006.252239] env[61243]: DEBUG nova.network.neutron [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Successfully updated port: ed08f6b9-7027-417b-91d5-e9639b3fed92 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1006.277167] env[61243]: DEBUG oslo_vmware.api [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339317, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.325658] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a0afbb7c-46b9-4395-9aa5-1b083bc9f11c tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "2c4217cd-af4a-4d64-a422-73955326df03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.344s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.490059] env[61243]: DEBUG nova.compute.manager [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1006.521249] env[61243]: DEBUG oslo_vmware.api [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339324, 'name': ReconfigVM_Task, 'duration_secs': 0.300863} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.521834] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Reconfigured VM instance instance-00000046 to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1006.527098] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6cab03e-75c6-4e41-b87c-398aa4d24c8d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.544778] env[61243]: DEBUG oslo_vmware.api [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1006.544778] env[61243]: value = "task-1339325" [ 1006.544778] env[61243]: _type = "Task" [ 1006.544778] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.553237] env[61243]: DEBUG oslo_vmware.api [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339325, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.674088] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339323, 'name': PowerOffVM_Task, 'duration_secs': 1.297355} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.674088] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1006.674349] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Updating instance '5660de64-7152-4832-960e-2deb1c0bc37e' progress to 17 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1006.754945] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "refresh_cache-dca3301d-ef3e-48ef-920c-866b4a086ea6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.755142] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquired lock "refresh_cache-dca3301d-ef3e-48ef-920c-866b4a086ea6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.755449] env[61243]: DEBUG nova.network.neutron [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1006.772697] env[61243]: ERROR nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [req-c00e1bfe-3f0a-42a0-8f5d-8171da3280ed] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c00e1bfe-3f0a-42a0-8f5d-8171da3280ed"}]} [ 1006.782971] env[61243]: DEBUG oslo_vmware.api [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339317, 'name': ReconfigVM_Task, 'duration_secs': 5.882063} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.783332] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.783528] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Reconfigured VM to detach interface {{(pid=61243) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1006.789253] env[61243]: DEBUG nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1006.802502] env[61243]: DEBUG nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1006.802726] env[61243]: DEBUG nova.compute.provider_tree [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1006.814418] env[61243]: DEBUG nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1006.832360] env[61243]: DEBUG nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1007.011345] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865f63d4-bd8d-4567-ac12-e3af4c57b76c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.015477] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.020458] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b0fb561-98e2-44bd-9340-1ddfeece4005 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.056651] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a07d64-9a43-474d-a609-5b524e179e3d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.063844] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.063844] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.063844] env[61243]: DEBUG nova.network.neutron [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1007.072798] env[61243]: DEBUG oslo_vmware.api [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339325, 'name': ReconfigVM_Task, 'duration_secs': 0.170114} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.075254] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285747', 'volume_id': 'bc3b5837-80e8-4d8a-8a68-5ec6c3b89880', 'name': 'volume-bc3b5837-80e8-4d8a-8a68-5ec6c3b89880', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'c1a0d242-4eae-4c03-8341-840b41341f17', 'attached_at': '', 'detached_at': '', 'volume_id': 'bc3b5837-80e8-4d8a-8a68-5ec6c3b89880', 'serial': 'bc3b5837-80e8-4d8a-8a68-5ec6c3b89880'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1007.079167] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22a8118-4a66-4c51-8b39-87bd79f58512 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.095794] env[61243]: DEBUG nova.compute.provider_tree [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1007.181466] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1007.181900] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1007.181900] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1007.182072] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1007.182244] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1007.182405] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1007.182619] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1007.182785] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1007.182958] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1007.183142] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1007.183353] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1007.188265] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-70aa0b16-ab28-46fc-8bcf-512c937f50f5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.203374] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1007.203374] env[61243]: value = "task-1339326" [ 1007.203374] env[61243]: _type = "Task" [ 1007.203374] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.213473] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339326, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.289528] env[61243]: DEBUG nova.network.neutron [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1007.430301] env[61243]: DEBUG nova.network.neutron [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Updating instance_info_cache with network_info: [{"id": "ed08f6b9-7027-417b-91d5-e9639b3fed92", "address": "fa:16:3e:8f:9c:75", "network": {"id": "523d4cf1-467c-4ea6-80a4-ebbd01016efa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-919325376-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7651b565d1647eb948fe07faba32a41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped08f6b9-70", "ovs_interfaceid": "ed08f6b9-7027-417b-91d5-e9639b3fed92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.575317] env[61243]: DEBUG nova.compute.manager [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Stashing vm_state: active {{(pid=61243) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1007.630303] env[61243]: DEBUG nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 134 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1007.630596] env[61243]: DEBUG nova.compute.provider_tree [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 134 to 135 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1007.630931] env[61243]: DEBUG nova.compute.provider_tree [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1007.643370] env[61243]: DEBUG nova.objects.instance [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lazy-loading 'flavor' on Instance uuid c1a0d242-4eae-4c03-8341-840b41341f17 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.718318] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339326, 'name': ReconfigVM_Task, 'duration_secs': 0.135187} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.718318] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Updating instance '5660de64-7152-4832-960e-2deb1c0bc37e' progress to 33 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1007.835497] env[61243]: INFO nova.network.neutron [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Port b91562c3-ce88-4ab3-8413-a5bfeb82aa48 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1007.835765] env[61243]: INFO nova.network.neutron [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Port 317b0929-bd21-499a-aeb7-39bb4b38b703 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1007.836201] env[61243]: DEBUG nova.network.neutron [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Updating instance_info_cache with network_info: [{"id": "2333ab37-7807-45d2-a74e-05404217df6d", "address": "fa:16:3e:ea:93:91", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2333ab37-78", "ovs_interfaceid": "2333ab37-7807-45d2-a74e-05404217df6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.933650] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Releasing lock "refresh_cache-dca3301d-ef3e-48ef-920c-866b4a086ea6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.933989] env[61243]: DEBUG nova.compute.manager [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Instance network_info: |[{"id": "ed08f6b9-7027-417b-91d5-e9639b3fed92", "address": "fa:16:3e:8f:9c:75", "network": {"id": "523d4cf1-467c-4ea6-80a4-ebbd01016efa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-919325376-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7651b565d1647eb948fe07faba32a41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped08f6b9-70", "ovs_interfaceid": "ed08f6b9-7027-417b-91d5-e9639b3fed92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1007.934456] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:9c:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '42f08482-a1da-405d-9918-d733d9f5173c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ed08f6b9-7027-417b-91d5-e9639b3fed92', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1007.941804] env[61243]: DEBUG oslo.service.loopingcall [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1007.942044] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1007.942291] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7d90ac2-a752-425c-8afa-dcfb4c524173 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.964048] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1007.964048] env[61243]: value = "task-1339327" [ 1007.964048] env[61243]: _type = "Task" [ 1007.964048] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.971782] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339327, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.095680] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.138736] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 4.541s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.141437] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.045s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.144559] env[61243]: INFO nova.compute.claims [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1008.161382] env[61243]: DEBUG nova.compute.manager [req-88fe109a-1e06-4980-b865-ef1fccf438d2 req-4ea001f4-b66b-494b-824b-8e2716861de3 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Received event network-changed-ed08f6b9-7027-417b-91d5-e9639b3fed92 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.161568] env[61243]: DEBUG nova.compute.manager [req-88fe109a-1e06-4980-b865-ef1fccf438d2 req-4ea001f4-b66b-494b-824b-8e2716861de3 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Refreshing instance network info cache due to event network-changed-ed08f6b9-7027-417b-91d5-e9639b3fed92. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1008.161838] env[61243]: DEBUG oslo_concurrency.lockutils [req-88fe109a-1e06-4980-b865-ef1fccf438d2 req-4ea001f4-b66b-494b-824b-8e2716861de3 service nova] Acquiring lock "refresh_cache-dca3301d-ef3e-48ef-920c-866b4a086ea6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.161974] env[61243]: DEBUG oslo_concurrency.lockutils [req-88fe109a-1e06-4980-b865-ef1fccf438d2 req-4ea001f4-b66b-494b-824b-8e2716861de3 service nova] Acquired lock "refresh_cache-dca3301d-ef3e-48ef-920c-866b4a086ea6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.162369] env[61243]: DEBUG nova.network.neutron [req-88fe109a-1e06-4980-b865-ef1fccf438d2 req-4ea001f4-b66b-494b-824b-8e2716861de3 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Refreshing network info cache for port ed08f6b9-7027-417b-91d5-e9639b3fed92 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1008.165699] env[61243]: INFO nova.scheduler.client.report [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleted allocations for instance 818476f0-e928-44d5-b3d4-101b7a05c9aa [ 1008.219479] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.219751] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.219907] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.220111] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.220292] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.223557] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1008.223776] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1008.223939] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1008.224143] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1008.224307] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1008.224451] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1008.224654] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1008.224820] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1008.224989] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1008.225182] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1008.225356] env[61243]: DEBUG nova.virt.hardware [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1008.230590] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Reconfiguring VM instance instance-0000005c to detach disk 2000 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1008.231090] env[61243]: INFO nova.compute.manager [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Terminating instance [ 1008.233053] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3afe3362-dc61-4752-9b1f-dbc05cafc7e6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.247251] env[61243]: DEBUG nova.compute.manager [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1008.247455] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1008.248500] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4845a005-8cb3-4237-9b72-a1f4a10b6962 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.257310] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1008.257552] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7eb8f3a9-eea1-474a-b73c-52a06266d410 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.259895] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1008.259895] env[61243]: value = "task-1339328" [ 1008.259895] env[61243]: _type = "Task" [ 1008.259895] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.264528] env[61243]: DEBUG oslo_vmware.api [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1008.264528] env[61243]: value = "task-1339329" [ 1008.264528] env[61243]: _type = "Task" [ 1008.264528] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.267978] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339328, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.275815] env[61243]: DEBUG oslo_vmware.api [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339329, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.338759] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "refresh_cache-b0b5b215-d7ff-4960-9d39-cf792d8d4656" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.473953] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339327, 'name': CreateVM_Task, 'duration_secs': 0.338725} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.474096] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1008.474776] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.474953] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.475305] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1008.475584] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b99e50ac-b938-4459-8c44-e75aa6098fe6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.480580] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1008.480580] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5283e04a-be5b-7491-facd-f40d10a220a5" [ 1008.480580] env[61243]: _type = "Task" [ 1008.480580] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.489059] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5283e04a-be5b-7491-facd-f40d10a220a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.657628] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f55f4b2c-9fc5-4f4e-94be-be62fc4f9ea5 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "c1a0d242-4eae-4c03-8341-840b41341f17" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.332s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.676221] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e6be0271-af57-45b6-aad7-34952d3ac368 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "818476f0-e928-44d5-b3d4-101b7a05c9aa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.229s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.774609] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339328, 'name': ReconfigVM_Task, 'duration_secs': 0.212527} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.775464] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Reconfigured VM instance instance-0000005c to detach disk 2000 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1008.776328] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ffa15e-9600-4f19-a3f1-5e8cc4867234 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.781724] env[61243]: DEBUG oslo_vmware.api [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339329, 'name': PowerOffVM_Task, 'duration_secs': 0.198412} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.782423] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1008.782703] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1008.783053] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-350e0593-6f20-4686-b1f1-aa50d3ac95c0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.805109] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 5660de64-7152-4832-960e-2deb1c0bc37e/5660de64-7152-4832-960e-2deb1c0bc37e.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1008.807020] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8fe02d4-b5b0-4f91-b8ea-bd46cad93e07 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.828763] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1008.828763] env[61243]: value = "task-1339331" [ 1008.828763] env[61243]: _type = "Task" [ 1008.828763] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.843121] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ddf220d-7ae9-43ce-8fc9-391779490ecf tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-b0b5b215-d7ff-4960-9d39-cf792d8d4656-b91562c3-ce88-4ab3-8413-a5bfeb82aa48" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 8.667s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.847784] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339331, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.907044] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1008.907452] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1008.907841] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Deleting the datastore file [datastore2] b0b5b215-d7ff-4960-9d39-cf792d8d4656 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1008.908079] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9ff371d7-10ab-449c-b54f-aac7349c3eb8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.914418] env[61243]: DEBUG oslo_vmware.api [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1008.914418] env[61243]: value = "task-1339332" [ 1008.914418] env[61243]: _type = "Task" [ 1008.914418] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.924974] env[61243]: DEBUG oslo_vmware.api [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339332, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.991885] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5283e04a-be5b-7491-facd-f40d10a220a5, 'name': SearchDatastore_Task, 'duration_secs': 0.036356} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.994528] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.994871] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1008.995210] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.995436] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.995690] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1008.996017] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85bbf0d2-2c16-447b-88dd-eb9b925c336a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.012788] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1009.013140] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1009.013898] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a54e854f-b4da-4316-bf19-ec6e7cfce334 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.019498] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1009.019498] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e6fa80-b454-748a-42b8-b9a6f65d3137" [ 1009.019498] env[61243]: _type = "Task" [ 1009.019498] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.027877] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e6fa80-b454-748a-42b8-b9a6f65d3137, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.105514] env[61243]: DEBUG nova.network.neutron [req-88fe109a-1e06-4980-b865-ef1fccf438d2 req-4ea001f4-b66b-494b-824b-8e2716861de3 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Updated VIF entry in instance network info cache for port ed08f6b9-7027-417b-91d5-e9639b3fed92. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1009.106045] env[61243]: DEBUG nova.network.neutron [req-88fe109a-1e06-4980-b865-ef1fccf438d2 req-4ea001f4-b66b-494b-824b-8e2716861de3 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Updating instance_info_cache with network_info: [{"id": "ed08f6b9-7027-417b-91d5-e9639b3fed92", "address": "fa:16:3e:8f:9c:75", "network": {"id": "523d4cf1-467c-4ea6-80a4-ebbd01016efa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-919325376-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7651b565d1647eb948fe07faba32a41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped08f6b9-70", "ovs_interfaceid": "ed08f6b9-7027-417b-91d5-e9639b3fed92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.325017] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a746c5-af48-4870-973a-031217f89187 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.335028] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f381d764-f169-431a-b9ce-660972de6845 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.341880] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339331, 'name': ReconfigVM_Task, 'duration_secs': 0.40072} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.364833] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 5660de64-7152-4832-960e-2deb1c0bc37e/5660de64-7152-4832-960e-2deb1c0bc37e.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1009.365152] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Updating instance '5660de64-7152-4832-960e-2deb1c0bc37e' progress to 50 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1009.368901] env[61243]: DEBUG oslo_concurrency.lockutils [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "7ca74bdc-c272-4b6e-a4dd-56d97144b197" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.368901] env[61243]: DEBUG oslo_concurrency.lockutils [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "7ca74bdc-c272-4b6e-a4dd-56d97144b197" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.369089] env[61243]: DEBUG oslo_concurrency.lockutils [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "7ca74bdc-c272-4b6e-a4dd-56d97144b197-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.369280] env[61243]: DEBUG oslo_concurrency.lockutils [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "7ca74bdc-c272-4b6e-a4dd-56d97144b197-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.369458] env[61243]: DEBUG oslo_concurrency.lockutils [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "7ca74bdc-c272-4b6e-a4dd-56d97144b197-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.371758] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a4e2859-c6ca-419d-b707-52de27038c2a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.374435] env[61243]: INFO nova.compute.manager [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Terminating instance [ 1009.376291] env[61243]: DEBUG nova.compute.manager [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1009.376489] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1009.377189] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946d4984-83e3-4823-b2ce-08974ff56edc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.384068] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8333024-4e32-4c0f-9624-a648c598f7a0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.389610] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1009.390159] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd4aa4bf-c063-4176-b664-a368249bbf44 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.401637] env[61243]: DEBUG nova.compute.provider_tree [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1009.406121] env[61243]: DEBUG oslo_vmware.api [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1009.406121] env[61243]: value = "task-1339333" [ 1009.406121] env[61243]: _type = "Task" [ 1009.406121] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.414297] env[61243]: DEBUG oslo_vmware.api [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339333, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.423359] env[61243]: DEBUG oslo_vmware.api [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339332, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.248481} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.423624] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.423803] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1009.423983] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1009.424181] env[61243]: INFO nova.compute.manager [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1009.424415] env[61243]: DEBUG oslo.service.loopingcall [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.424605] env[61243]: DEBUG nova.compute.manager [-] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1009.424729] env[61243]: DEBUG nova.network.neutron [-] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1009.469801] env[61243]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 317b0929-bd21-499a-aeb7-39bb4b38b703 could not be found.", "detail": ""}} {{(pid=61243) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 1009.470060] env[61243]: DEBUG nova.network.neutron [-] Unable to show port 317b0929-bd21-499a-aeb7-39bb4b38b703 as it no longer exists. {{(pid=61243) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 1009.530614] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e6fa80-b454-748a-42b8-b9a6f65d3137, 'name': SearchDatastore_Task, 'duration_secs': 0.01178} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.531515] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f510652d-8844-4187-a015-7465b7ad9900 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.536862] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1009.536862] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52134805-1647-016b-23bf-6568b0cb8722" [ 1009.536862] env[61243]: _type = "Task" [ 1009.536862] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.544861] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52134805-1647-016b-23bf-6568b0cb8722, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.610810] env[61243]: DEBUG oslo_concurrency.lockutils [req-88fe109a-1e06-4980-b865-ef1fccf438d2 req-4ea001f4-b66b-494b-824b-8e2716861de3 service nova] Releasing lock "refresh_cache-dca3301d-ef3e-48ef-920c-866b4a086ea6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.611109] env[61243]: DEBUG nova.compute.manager [req-88fe109a-1e06-4980-b865-ef1fccf438d2 req-4ea001f4-b66b-494b-824b-8e2716861de3 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Received event network-vif-deleted-b91562c3-ce88-4ab3-8413-a5bfeb82aa48 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.611318] env[61243]: DEBUG nova.compute.manager [req-88fe109a-1e06-4980-b865-ef1fccf438d2 req-4ea001f4-b66b-494b-824b-8e2716861de3 service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Received event network-vif-deleted-317b0929-bd21-499a-aeb7-39bb4b38b703 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.632098] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "c1a0d242-4eae-4c03-8341-840b41341f17" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.632356] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "c1a0d242-4eae-4c03-8341-840b41341f17" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.632575] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "c1a0d242-4eae-4c03-8341-840b41341f17-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.632769] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "c1a0d242-4eae-4c03-8341-840b41341f17-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.632995] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "c1a0d242-4eae-4c03-8341-840b41341f17-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.635281] env[61243]: INFO nova.compute.manager [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Terminating instance [ 1009.637169] env[61243]: DEBUG nova.compute.manager [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1009.637395] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1009.638256] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba0eb6b9-9b07-40df-8de0-505720d89f06 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.646317] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1009.646317] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7cf58b88-9675-445b-8458-6a958a42a7e9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.652240] env[61243]: DEBUG oslo_vmware.api [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1009.652240] env[61243]: value = "task-1339334" [ 1009.652240] env[61243]: _type = "Task" [ 1009.652240] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.662634] env[61243]: DEBUG oslo_vmware.api [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339334, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.873736] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76c22766-1304-4125-95a2-003bb497b695 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.893643] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d670be-02de-4633-8cf3-c7e2788d0583 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.913415] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Updating instance '5660de64-7152-4832-960e-2deb1c0bc37e' progress to 67 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1009.924668] env[61243]: DEBUG oslo_vmware.api [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339333, 'name': PowerOffVM_Task, 'duration_secs': 0.186209} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.925031] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1009.925216] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1009.925475] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e942a62-7ad0-4cc6-8066-795f66a05d76 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.943851] env[61243]: DEBUG nova.scheduler.client.report [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 135 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1009.944202] env[61243]: DEBUG nova.compute.provider_tree [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 135 to 136 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1009.944415] env[61243]: DEBUG nova.compute.provider_tree [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1010.005086] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1010.005327] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1010.005524] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleting the datastore file [datastore2] 7ca74bdc-c272-4b6e-a4dd-56d97144b197 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.005836] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3e04f28a-4a59-44ff-805f-bcad49bdb7a9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.011426] env[61243]: DEBUG oslo_vmware.api [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1010.011426] env[61243]: value = "task-1339336" [ 1010.011426] env[61243]: _type = "Task" [ 1010.011426] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.019073] env[61243]: DEBUG oslo_vmware.api [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339336, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.048021] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52134805-1647-016b-23bf-6568b0cb8722, 'name': SearchDatastore_Task, 'duration_secs': 0.012163} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.048021] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.048021] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] dca3301d-ef3e-48ef-920c-866b4a086ea6/dca3301d-ef3e-48ef-920c-866b4a086ea6.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1010.048021] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3e871f1e-b7c0-4255-bccb-a7dc29f7d179 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.052196] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1010.052196] env[61243]: value = "task-1339337" [ 1010.052196] env[61243]: _type = "Task" [ 1010.052196] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.059430] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339337, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.161593] env[61243]: DEBUG oslo_vmware.api [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339334, 'name': PowerOffVM_Task, 'duration_secs': 0.342857} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.161866] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1010.162391] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1010.162391] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-090beec5-984d-4cbb-a2e2-1f524cade5b5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.189279] env[61243]: DEBUG nova.compute.manager [req-24605aaa-a54c-4eb9-b90d-a48286764743 req-342e50e8-e6f1-41dc-abda-1d65ef06782d service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Received event network-vif-deleted-2333ab37-7807-45d2-a74e-05404217df6d {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.189279] env[61243]: INFO nova.compute.manager [req-24605aaa-a54c-4eb9-b90d-a48286764743 req-342e50e8-e6f1-41dc-abda-1d65ef06782d service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Neutron deleted interface 2333ab37-7807-45d2-a74e-05404217df6d; detaching it from the instance and deleting it from the info cache [ 1010.189279] env[61243]: DEBUG nova.network.neutron [req-24605aaa-a54c-4eb9-b90d-a48286764743 req-342e50e8-e6f1-41dc-abda-1d65ef06782d service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.229108] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1010.229387] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1010.229575] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Deleting the datastore file [datastore2] c1a0d242-4eae-4c03-8341-840b41341f17 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.229853] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7cd88d37-8c06-4249-bd8c-04d00b8f4f6b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.236546] env[61243]: DEBUG oslo_vmware.api [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1010.236546] env[61243]: value = "task-1339339" [ 1010.236546] env[61243]: _type = "Task" [ 1010.236546] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.244980] env[61243]: DEBUG oslo_vmware.api [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339339, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.356035] env[61243]: DEBUG nova.network.neutron [-] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.450368] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.450872] env[61243]: DEBUG nova.compute.manager [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1010.454471] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.439s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.457266] env[61243]: INFO nova.compute.claims [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1010.483152] env[61243]: DEBUG nova.network.neutron [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Port 2b065192-815c-492f-8f79-bfa9553ff080 binding to destination host cpu-1 is already ACTIVE {{(pid=61243) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1010.524099] env[61243]: DEBUG oslo_vmware.api [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339336, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186467} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.524499] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1010.524673] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1010.524892] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1010.525132] env[61243]: INFO nova.compute.manager [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1010.525453] env[61243]: DEBUG oslo.service.loopingcall [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1010.525673] env[61243]: DEBUG nova.compute.manager [-] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1010.525784] env[61243]: DEBUG nova.network.neutron [-] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1010.562809] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339337, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.692024] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5326883a-04aa-4bcb-afa4-80ac6b93f4f6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.702734] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2d1556-c75e-4a0d-9c72-e2e34dfa601d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.733890] env[61243]: DEBUG nova.compute.manager [req-24605aaa-a54c-4eb9-b90d-a48286764743 req-342e50e8-e6f1-41dc-abda-1d65ef06782d service nova] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Detach interface failed, port_id=2333ab37-7807-45d2-a74e-05404217df6d, reason: Instance b0b5b215-d7ff-4960-9d39-cf792d8d4656 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1010.746909] env[61243]: DEBUG oslo_vmware.api [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339339, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.859142] env[61243]: INFO nova.compute.manager [-] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Took 1.43 seconds to deallocate network for instance. [ 1010.956456] env[61243]: DEBUG nova.compute.utils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1010.957855] env[61243]: DEBUG nova.compute.manager [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1010.958097] env[61243]: DEBUG nova.network.neutron [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1010.997211] env[61243]: DEBUG nova.policy [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e027c639bf9c4157b6d135bc5526e5aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '92a24c66820042a48e8590532dc36908', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1011.063080] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339337, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522316} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.063410] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] dca3301d-ef3e-48ef-920c-866b4a086ea6/dca3301d-ef3e-48ef-920c-866b4a086ea6.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1011.063668] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1011.063931] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf301c90-7ec3-4a49-aa66-581f5813affb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.070631] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1011.070631] env[61243]: value = "task-1339340" [ 1011.070631] env[61243]: _type = "Task" [ 1011.070631] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.079465] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339340, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.248795] env[61243]: DEBUG oslo_vmware.api [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339339, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.56453} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.249095] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.249295] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1011.249488] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1011.249672] env[61243]: INFO nova.compute.manager [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Took 1.61 seconds to destroy the instance on the hypervisor. [ 1011.249918] env[61243]: DEBUG oslo.service.loopingcall [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.250133] env[61243]: DEBUG nova.compute.manager [-] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1011.250233] env[61243]: DEBUG nova.network.neutron [-] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1011.270364] env[61243]: DEBUG nova.network.neutron [-] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.282341] env[61243]: DEBUG nova.network.neutron [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Successfully created port: b3d60ea2-8847-41fc-aa33-1998bb27a115 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1011.366025] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.461121] env[61243]: DEBUG nova.compute.manager [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1011.518644] env[61243]: DEBUG oslo_concurrency.lockutils [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "5660de64-7152-4832-960e-2deb1c0bc37e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.518878] env[61243]: DEBUG oslo_concurrency.lockutils [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "5660de64-7152-4832-960e-2deb1c0bc37e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.519074] env[61243]: DEBUG oslo_concurrency.lockutils [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "5660de64-7152-4832-960e-2deb1c0bc37e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.590314] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339340, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062638} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.590768] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1011.591811] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7901905-a487-49b9-8f2c-1ee0127302c4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.618171] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] dca3301d-ef3e-48ef-920c-866b4a086ea6/dca3301d-ef3e-48ef-920c-866b4a086ea6.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1011.621573] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e90d3c9-8be4-41aa-bd63-d317bf0dcee1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.641954] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1011.641954] env[61243]: value = "task-1339341" [ 1011.641954] env[61243]: _type = "Task" [ 1011.641954] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.651243] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339341, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.717118] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e292359-1b96-4d99-b672-39f35d83445d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.725016] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96554dc3-567a-4c20-a915-9c3543f66346 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.760444] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d33d6570-0292-41ad-af87-8bd960ff90c7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.767792] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b877fbd-966f-496e-8a57-6af430dc7eb2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.772227] env[61243]: INFO nova.compute.manager [-] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Took 1.25 seconds to deallocate network for instance. [ 1011.782240] env[61243]: DEBUG nova.compute.provider_tree [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.156484] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339341, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.218789] env[61243]: DEBUG nova.compute.manager [req-d113f167-5621-4127-809f-1429da3042d0 req-d331c3a3-44a2-4b9b-b362-62af8c4bcbbd service nova] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Received event network-vif-deleted-1dbcd762-9bfc-4f3e-86d9-d9452a88349b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.218789] env[61243]: DEBUG nova.compute.manager [req-d113f167-5621-4127-809f-1429da3042d0 req-d331c3a3-44a2-4b9b-b362-62af8c4bcbbd service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Received event network-vif-deleted-970e8509-164d-4ddf-8f0c-795e92883f6b {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.218927] env[61243]: INFO nova.compute.manager [req-d113f167-5621-4127-809f-1429da3042d0 req-d331c3a3-44a2-4b9b-b362-62af8c4bcbbd service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Neutron deleted interface 970e8509-164d-4ddf-8f0c-795e92883f6b; detaching it from the instance and deleting it from the info cache [ 1012.219094] env[61243]: DEBUG nova.network.neutron [req-d113f167-5621-4127-809f-1429da3042d0 req-d331c3a3-44a2-4b9b-b362-62af8c4bcbbd service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.288618] env[61243]: DEBUG nova.scheduler.client.report [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.292579] env[61243]: DEBUG oslo_concurrency.lockutils [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.392057] env[61243]: DEBUG nova.network.neutron [-] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.470641] env[61243]: DEBUG nova.compute.manager [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1012.497959] env[61243]: DEBUG nova.virt.hardware [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1012.498236] env[61243]: DEBUG nova.virt.hardware [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1012.498403] env[61243]: DEBUG nova.virt.hardware [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1012.499039] env[61243]: DEBUG nova.virt.hardware [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1012.499039] env[61243]: DEBUG nova.virt.hardware [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1012.499039] env[61243]: DEBUG nova.virt.hardware [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1012.499265] env[61243]: DEBUG nova.virt.hardware [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1012.499359] env[61243]: DEBUG nova.virt.hardware [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1012.499540] env[61243]: DEBUG nova.virt.hardware [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1012.499712] env[61243]: DEBUG nova.virt.hardware [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1012.499892] env[61243]: DEBUG nova.virt.hardware [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1012.500788] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aabeeef-e746-4897-b4ba-028861e5555a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.510811] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7580ed9-0e4f-44b0-a88a-69cbedc2e933 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.557681] env[61243]: DEBUG oslo_concurrency.lockutils [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "refresh_cache-5660de64-7152-4832-960e-2deb1c0bc37e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.557884] env[61243]: DEBUG oslo_concurrency.lockutils [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "refresh_cache-5660de64-7152-4832-960e-2deb1c0bc37e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.558081] env[61243]: DEBUG nova.network.neutron [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1012.652792] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339341, 'name': ReconfigVM_Task, 'duration_secs': 0.686543} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.653104] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Reconfigured VM instance instance-00000060 to attach disk [datastore2] dca3301d-ef3e-48ef-920c-866b4a086ea6/dca3301d-ef3e-48ef-920c-866b4a086ea6.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1012.653788] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6b1e7553-1ca5-4f52-8f45-3630208ebdbe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.662668] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1012.662668] env[61243]: value = "task-1339342" [ 1012.662668] env[61243]: _type = "Task" [ 1012.662668] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.670315] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339342, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.721804] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9fd9bb50-e955-4ea9-b521-ad027bc1038e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.730744] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-178e5b09-0bec-4b6d-a43d-34c4f760996e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.756927] env[61243]: DEBUG nova.compute.manager [req-d113f167-5621-4127-809f-1429da3042d0 req-d331c3a3-44a2-4b9b-b362-62af8c4bcbbd service nova] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Detach interface failed, port_id=970e8509-164d-4ddf-8f0c-795e92883f6b, reason: Instance c1a0d242-4eae-4c03-8341-840b41341f17 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1012.794288] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.794949] env[61243]: DEBUG nova.compute.manager [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1012.797654] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 4.702s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.894693] env[61243]: INFO nova.compute.manager [-] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Took 1.64 seconds to deallocate network for instance. [ 1012.962324] env[61243]: DEBUG nova.network.neutron [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Successfully updated port: b3d60ea2-8847-41fc-aa33-1998bb27a115 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1013.175233] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339342, 'name': Rename_Task, 'duration_secs': 0.133215} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.175398] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1013.175642] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-82700551-ae06-4c1b-a3a2-7c0a62c844a0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.182050] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1013.182050] env[61243]: value = "task-1339343" [ 1013.182050] env[61243]: _type = "Task" [ 1013.182050] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.189433] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339343, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.272043] env[61243]: DEBUG nova.network.neutron [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Updating instance_info_cache with network_info: [{"id": "2b065192-815c-492f-8f79-bfa9553ff080", "address": "fa:16:3e:94:3e:9e", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b065192-81", "ovs_interfaceid": "2b065192-815c-492f-8f79-bfa9553ff080", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.301207] env[61243]: DEBUG nova.compute.utils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1013.304907] env[61243]: INFO nova.compute.claims [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1013.308507] env[61243]: DEBUG nova.compute.manager [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1013.308745] env[61243]: DEBUG nova.network.neutron [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1013.350697] env[61243]: DEBUG nova.policy [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ecfff92a554145c0bd024e59fb3d6f52', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66e2ba61b3354ef1b496232264d27600', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1013.400581] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.468197] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Acquiring lock "refresh_cache-0e03eee3-e7bc-483f-948f-a5b408375e24" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.468396] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Acquired lock "refresh_cache-0e03eee3-e7bc-483f-948f-a5b408375e24" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.468483] env[61243]: DEBUG nova.network.neutron [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1013.616465] env[61243]: DEBUG nova.network.neutron [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Successfully created port: f2a48425-d436-40c9-a2dc-694e6313da6d {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1013.693874] env[61243]: DEBUG oslo_vmware.api [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339343, 'name': PowerOnVM_Task, 'duration_secs': 0.491401} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.694167] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1013.694381] env[61243]: INFO nova.compute.manager [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Took 8.07 seconds to spawn the instance on the hypervisor. [ 1013.694588] env[61243]: DEBUG nova.compute.manager [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1013.695362] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e91bb6-7de1-424f-86c0-f2858249ed4e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.775562] env[61243]: DEBUG oslo_concurrency.lockutils [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "refresh_cache-5660de64-7152-4832-960e-2deb1c0bc37e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.810932] env[61243]: DEBUG nova.compute.manager [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1013.816764] env[61243]: INFO nova.compute.resource_tracker [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Updating resource usage from migration 8403be2c-7c0b-4e7f-b48e-ad74b99819e9 [ 1013.975190] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe679e57-1841-4227-9db8-9026d2457473 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.983952] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a64360fc-3f15-481f-a4d4-da330d5d8d03 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.015535] env[61243]: DEBUG nova.network.neutron [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1014.017783] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b64dbd-5d78-4527-b03d-7a07fc70e270 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.025045] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba77847-5797-437e-a8f7-3ba48769afa4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.037725] env[61243]: DEBUG nova.compute.provider_tree [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.166704] env[61243]: DEBUG nova.network.neutron [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Updating instance_info_cache with network_info: [{"id": "b3d60ea2-8847-41fc-aa33-1998bb27a115", "address": "fa:16:3e:73:1c:43", "network": {"id": "642f7241-c6f9-4fa4-8bec-76c723f26062", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-437990091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92a24c66820042a48e8590532dc36908", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3d60ea2-88", "ovs_interfaceid": "b3d60ea2-8847-41fc-aa33-1998bb27a115", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.211690] env[61243]: INFO nova.compute.manager [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Took 17.43 seconds to build instance. [ 1014.249951] env[61243]: DEBUG nova.compute.manager [req-2eb149fe-2589-4c6f-9858-21c038841fc7 req-df9de37f-ba85-4f94-9e66-2165674e4a3d service nova] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Received event network-vif-plugged-b3d60ea2-8847-41fc-aa33-1998bb27a115 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.250209] env[61243]: DEBUG oslo_concurrency.lockutils [req-2eb149fe-2589-4c6f-9858-21c038841fc7 req-df9de37f-ba85-4f94-9e66-2165674e4a3d service nova] Acquiring lock "0e03eee3-e7bc-483f-948f-a5b408375e24-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.250436] env[61243]: DEBUG oslo_concurrency.lockutils [req-2eb149fe-2589-4c6f-9858-21c038841fc7 req-df9de37f-ba85-4f94-9e66-2165674e4a3d service nova] Lock "0e03eee3-e7bc-483f-948f-a5b408375e24-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.250618] env[61243]: DEBUG oslo_concurrency.lockutils [req-2eb149fe-2589-4c6f-9858-21c038841fc7 req-df9de37f-ba85-4f94-9e66-2165674e4a3d service nova] Lock "0e03eee3-e7bc-483f-948f-a5b408375e24-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.250793] env[61243]: DEBUG nova.compute.manager [req-2eb149fe-2589-4c6f-9858-21c038841fc7 req-df9de37f-ba85-4f94-9e66-2165674e4a3d service nova] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] No waiting events found dispatching network-vif-plugged-b3d60ea2-8847-41fc-aa33-1998bb27a115 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1014.250966] env[61243]: WARNING nova.compute.manager [req-2eb149fe-2589-4c6f-9858-21c038841fc7 req-df9de37f-ba85-4f94-9e66-2165674e4a3d service nova] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Received unexpected event network-vif-plugged-b3d60ea2-8847-41fc-aa33-1998bb27a115 for instance with vm_state building and task_state spawning. [ 1014.251147] env[61243]: DEBUG nova.compute.manager [req-2eb149fe-2589-4c6f-9858-21c038841fc7 req-df9de37f-ba85-4f94-9e66-2165674e4a3d service nova] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Received event network-changed-b3d60ea2-8847-41fc-aa33-1998bb27a115 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.251306] env[61243]: DEBUG nova.compute.manager [req-2eb149fe-2589-4c6f-9858-21c038841fc7 req-df9de37f-ba85-4f94-9e66-2165674e4a3d service nova] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Refreshing instance network info cache due to event network-changed-b3d60ea2-8847-41fc-aa33-1998bb27a115. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1014.251480] env[61243]: DEBUG oslo_concurrency.lockutils [req-2eb149fe-2589-4c6f-9858-21c038841fc7 req-df9de37f-ba85-4f94-9e66-2165674e4a3d service nova] Acquiring lock "refresh_cache-0e03eee3-e7bc-483f-948f-a5b408375e24" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.302333] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e49d99-cffc-463c-9d05-f11a8d95c58f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.325485] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f5ba9b-471b-4af3-9a41-68829e82d43f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.332967] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Updating instance '5660de64-7152-4832-960e-2deb1c0bc37e' progress to 83 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1014.542011] env[61243]: DEBUG nova.scheduler.client.report [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.669532] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Releasing lock "refresh_cache-0e03eee3-e7bc-483f-948f-a5b408375e24" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.669824] env[61243]: DEBUG nova.compute.manager [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Instance network_info: |[{"id": "b3d60ea2-8847-41fc-aa33-1998bb27a115", "address": "fa:16:3e:73:1c:43", "network": {"id": "642f7241-c6f9-4fa4-8bec-76c723f26062", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-437990091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92a24c66820042a48e8590532dc36908", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3d60ea2-88", "ovs_interfaceid": "b3d60ea2-8847-41fc-aa33-1998bb27a115", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1014.670170] env[61243]: DEBUG oslo_concurrency.lockutils [req-2eb149fe-2589-4c6f-9858-21c038841fc7 req-df9de37f-ba85-4f94-9e66-2165674e4a3d service nova] Acquired lock "refresh_cache-0e03eee3-e7bc-483f-948f-a5b408375e24" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.670369] env[61243]: DEBUG nova.network.neutron [req-2eb149fe-2589-4c6f-9858-21c038841fc7 req-df9de37f-ba85-4f94-9e66-2165674e4a3d service nova] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Refreshing network info cache for port b3d60ea2-8847-41fc-aa33-1998bb27a115 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1014.671663] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:1c:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60bdba1a-14cf-46b2-9d8b-aeaf4d80c815', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b3d60ea2-8847-41fc-aa33-1998bb27a115', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1014.678915] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Creating folder: Project (92a24c66820042a48e8590532dc36908). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1014.682138] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-77b5bd37-9501-4662-9000-6f71f5b16ee6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.692787] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Created folder: Project (92a24c66820042a48e8590532dc36908) in parent group-v285636. [ 1014.692993] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Creating folder: Instances. Parent ref: group-v285770. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1014.693249] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d1bec38-f23b-474f-8265-b9c18643d90f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.701798] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Created folder: Instances in parent group-v285770. [ 1014.702054] env[61243]: DEBUG oslo.service.loopingcall [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.702250] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1014.702455] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f8ef239a-f4d6-493e-b56f-55beb46f7e61 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.718571] env[61243]: DEBUG oslo_concurrency.lockutils [None req-e987970a-a793-439c-b2e0-bfd2ed8e22be tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "dca3301d-ef3e-48ef-920c-866b4a086ea6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.947s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.722790] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1014.722790] env[61243]: value = "task-1339346" [ 1014.722790] env[61243]: _type = "Task" [ 1014.722790] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.730361] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339346, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.826708] env[61243]: DEBUG nova.compute.manager [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1014.839065] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1014.839519] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1203feb-9047-47af-b01b-aaf6e643036a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.846483] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1014.846483] env[61243]: value = "task-1339347" [ 1014.846483] env[61243]: _type = "Task" [ 1014.846483] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.857977] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339347, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.860197] env[61243]: DEBUG nova.virt.hardware [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1014.860460] env[61243]: DEBUG nova.virt.hardware [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1014.860631] env[61243]: DEBUG nova.virt.hardware [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1014.860829] env[61243]: DEBUG nova.virt.hardware [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1014.860985] env[61243]: DEBUG nova.virt.hardware [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1014.861161] env[61243]: DEBUG nova.virt.hardware [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1014.861387] env[61243]: DEBUG nova.virt.hardware [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1014.861587] env[61243]: DEBUG nova.virt.hardware [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1014.861777] env[61243]: DEBUG nova.virt.hardware [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1014.861952] env[61243]: DEBUG nova.virt.hardware [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1014.862149] env[61243]: DEBUG nova.virt.hardware [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1014.862964] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd7f552-f8f1-4b7e-acc3-2e142f05fabf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.872499] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a549cdb-6507-49ba-a390-fb1842ef341f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.939133] env[61243]: DEBUG nova.network.neutron [req-2eb149fe-2589-4c6f-9858-21c038841fc7 req-df9de37f-ba85-4f94-9e66-2165674e4a3d service nova] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Updated VIF entry in instance network info cache for port b3d60ea2-8847-41fc-aa33-1998bb27a115. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1014.939241] env[61243]: DEBUG nova.network.neutron [req-2eb149fe-2589-4c6f-9858-21c038841fc7 req-df9de37f-ba85-4f94-9e66-2165674e4a3d service nova] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Updating instance_info_cache with network_info: [{"id": "b3d60ea2-8847-41fc-aa33-1998bb27a115", "address": "fa:16:3e:73:1c:43", "network": {"id": "642f7241-c6f9-4fa4-8bec-76c723f26062", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-437990091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92a24c66820042a48e8590532dc36908", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3d60ea2-88", "ovs_interfaceid": "b3d60ea2-8847-41fc-aa33-1998bb27a115", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.050022] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.250s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.050022] env[61243]: INFO nova.compute.manager [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Migrating [ 1015.057588] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.692s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.058135] env[61243]: DEBUG nova.objects.instance [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lazy-loading 'resources' on Instance uuid b0b5b215-d7ff-4960-9d39-cf792d8d4656 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.188410] env[61243]: DEBUG nova.network.neutron [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Successfully updated port: f2a48425-d436-40c9-a2dc-694e6313da6d {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1015.233021] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339346, 'name': CreateVM_Task, 'duration_secs': 0.388887} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.233267] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1015.233977] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.234178] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.234514] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1015.234783] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c977095c-7746-4ab6-9f7c-bc9ae846305f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.239581] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Waiting for the task: (returnval){ [ 1015.239581] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52065823-187f-d810-31e7-5c2106fbcc22" [ 1015.239581] env[61243]: _type = "Task" [ 1015.239581] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.247167] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52065823-187f-d810-31e7-5c2106fbcc22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.359638] env[61243]: DEBUG oslo_vmware.api [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339347, 'name': PowerOnVM_Task, 'duration_secs': 0.371548} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.359900] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1015.360103] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-27c845f1-8e84-400d-b08c-fda5c3595a67 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Updating instance '5660de64-7152-4832-960e-2deb1c0bc37e' progress to 100 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1015.442179] env[61243]: DEBUG oslo_concurrency.lockutils [req-2eb149fe-2589-4c6f-9858-21c038841fc7 req-df9de37f-ba85-4f94-9e66-2165674e4a3d service nova] Releasing lock "refresh_cache-0e03eee3-e7bc-483f-948f-a5b408375e24" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.578255] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "refresh_cache-2c4217cd-af4a-4d64-a422-73955326df03" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.578590] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "refresh_cache-2c4217cd-af4a-4d64-a422-73955326df03" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.578627] env[61243]: DEBUG nova.network.neutron [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1015.691491] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.691645] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.691799] env[61243]: DEBUG nova.network.neutron [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1015.750781] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52065823-187f-d810-31e7-5c2106fbcc22, 'name': SearchDatastore_Task, 'duration_secs': 0.010197} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.751096] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.751356] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1015.751618] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.751784] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.752189] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1015.752260] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d0abccc5-c0c0-4a06-bd08-33a9a44e860e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.755638] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e68c8d4-2932-49a9-9ac7-757ebfcb3031 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.761960] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1015.762163] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1015.762860] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4458aeac-48a2-42c5-91b3-18a29e70f054 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.765734] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-774c1823-b67a-4dbf-9749-56b8ec5f567e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.773117] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Waiting for the task: (returnval){ [ 1015.773117] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e8d648-d18d-20a0-9268-94db828b045d" [ 1015.773117] env[61243]: _type = "Task" [ 1015.773117] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.802247] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b94646-6bf5-44aa-9d75-0924a6c66a0d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.811401] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e8d648-d18d-20a0-9268-94db828b045d, 'name': SearchDatastore_Task, 'duration_secs': 0.015178} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.816158] env[61243]: DEBUG nova.compute.manager [req-ee07792f-7845-4425-83ec-e5153f56bcfa req-236c5f89-e968-4acd-a363-aa41cf291cf3 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Received event network-changed-ed08f6b9-7027-417b-91d5-e9639b3fed92 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.816373] env[61243]: DEBUG nova.compute.manager [req-ee07792f-7845-4425-83ec-e5153f56bcfa req-236c5f89-e968-4acd-a363-aa41cf291cf3 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Refreshing instance network info cache due to event network-changed-ed08f6b9-7027-417b-91d5-e9639b3fed92. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1015.816626] env[61243]: DEBUG oslo_concurrency.lockutils [req-ee07792f-7845-4425-83ec-e5153f56bcfa req-236c5f89-e968-4acd-a363-aa41cf291cf3 service nova] Acquiring lock "refresh_cache-dca3301d-ef3e-48ef-920c-866b4a086ea6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.816777] env[61243]: DEBUG oslo_concurrency.lockutils [req-ee07792f-7845-4425-83ec-e5153f56bcfa req-236c5f89-e968-4acd-a363-aa41cf291cf3 service nova] Acquired lock "refresh_cache-dca3301d-ef3e-48ef-920c-866b4a086ea6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.816948] env[61243]: DEBUG nova.network.neutron [req-ee07792f-7845-4425-83ec-e5153f56bcfa req-236c5f89-e968-4acd-a363-aa41cf291cf3 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Refreshing network info cache for port ed08f6b9-7027-417b-91d5-e9639b3fed92 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1015.818667] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84d97119-e396-43c2-ae18-6e26a1e134f6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.821992] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54681b94-8f5f-49c4-a604-539ef12f7b0c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.830153] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Waiting for the task: (returnval){ [ 1015.830153] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5241bacd-b48a-ab9f-df41-28d8f7fc8886" [ 1015.830153] env[61243]: _type = "Task" [ 1015.830153] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.839699] env[61243]: DEBUG nova.compute.provider_tree [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.852419] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5241bacd-b48a-ab9f-df41-28d8f7fc8886, 'name': SearchDatastore_Task, 'duration_secs': 0.017152} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.853596] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.854053] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 0e03eee3-e7bc-483f-948f-a5b408375e24/0e03eee3-e7bc-483f-948f-a5b408375e24.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1015.854404] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-94941260-8b49-437f-b7fc-b3a0348a2c6f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.862902] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Waiting for the task: (returnval){ [ 1015.862902] env[61243]: value = "task-1339348" [ 1015.862902] env[61243]: _type = "Task" [ 1015.862902] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.878070] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': task-1339348, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.245457] env[61243]: DEBUG nova.network.neutron [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1016.278372] env[61243]: DEBUG nova.compute.manager [req-14312bca-422d-40c4-ad8f-86da87da2fe4 req-be0b42bf-752a-4802-8139-7dce513b7cd7 service nova] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Received event network-vif-plugged-f2a48425-d436-40c9-a2dc-694e6313da6d {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1016.278618] env[61243]: DEBUG oslo_concurrency.lockutils [req-14312bca-422d-40c4-ad8f-86da87da2fe4 req-be0b42bf-752a-4802-8139-7dce513b7cd7 service nova] Acquiring lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.278838] env[61243]: DEBUG oslo_concurrency.lockutils [req-14312bca-422d-40c4-ad8f-86da87da2fe4 req-be0b42bf-752a-4802-8139-7dce513b7cd7 service nova] Lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.279017] env[61243]: DEBUG oslo_concurrency.lockutils [req-14312bca-422d-40c4-ad8f-86da87da2fe4 req-be0b42bf-752a-4802-8139-7dce513b7cd7 service nova] Lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.279275] env[61243]: DEBUG nova.compute.manager [req-14312bca-422d-40c4-ad8f-86da87da2fe4 req-be0b42bf-752a-4802-8139-7dce513b7cd7 service nova] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] No waiting events found dispatching network-vif-plugged-f2a48425-d436-40c9-a2dc-694e6313da6d {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1016.279455] env[61243]: WARNING nova.compute.manager [req-14312bca-422d-40c4-ad8f-86da87da2fe4 req-be0b42bf-752a-4802-8139-7dce513b7cd7 service nova] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Received unexpected event network-vif-plugged-f2a48425-d436-40c9-a2dc-694e6313da6d for instance with vm_state building and task_state spawning. [ 1016.279627] env[61243]: DEBUG nova.compute.manager [req-14312bca-422d-40c4-ad8f-86da87da2fe4 req-be0b42bf-752a-4802-8139-7dce513b7cd7 service nova] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Received event network-changed-f2a48425-d436-40c9-a2dc-694e6313da6d {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1016.279795] env[61243]: DEBUG nova.compute.manager [req-14312bca-422d-40c4-ad8f-86da87da2fe4 req-be0b42bf-752a-4802-8139-7dce513b7cd7 service nova] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Refreshing instance network info cache due to event network-changed-f2a48425-d436-40c9-a2dc-694e6313da6d. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1016.279948] env[61243]: DEBUG oslo_concurrency.lockutils [req-14312bca-422d-40c4-ad8f-86da87da2fe4 req-be0b42bf-752a-4802-8139-7dce513b7cd7 service nova] Acquiring lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.346357] env[61243]: DEBUG nova.scheduler.client.report [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1016.378199] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': task-1339348, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49971} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.378661] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 0e03eee3-e7bc-483f-948f-a5b408375e24/0e03eee3-e7bc-483f-948f-a5b408375e24.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1016.378890] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1016.379160] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d06f3286-13ff-47aa-ae51-823aebd46eb2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.389495] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Waiting for the task: (returnval){ [ 1016.389495] env[61243]: value = "task-1339349" [ 1016.389495] env[61243]: _type = "Task" [ 1016.389495] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.397645] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': task-1339349, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.433425] env[61243]: DEBUG nova.network.neutron [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Updating instance_info_cache with network_info: [{"id": "611257cf-50e1-491a-beae-8c942398013d", "address": "fa:16:3e:6a:c8:a1", "network": {"id": "0788f9ac-5a15-4431-a711-9b4f6cf79fbc", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1210957657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "113713170fb94b8a8d7985b9a02faf77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c7821ea-f92f-4f06-a4cb-05e1186a9d22", "external-id": "nsx-vlan-transportzone-69", "segmentation_id": 69, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap611257cf-50", "ovs_interfaceid": "611257cf-50e1-491a-beae-8c942398013d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.523458] env[61243]: DEBUG nova.network.neutron [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating instance_info_cache with network_info: [{"id": "f2a48425-d436-40c9-a2dc-694e6313da6d", "address": "fa:16:3e:34:84:e2", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2a48425-d4", "ovs_interfaceid": "f2a48425-d436-40c9-a2dc-694e6313da6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.637206] env[61243]: DEBUG nova.network.neutron [req-ee07792f-7845-4425-83ec-e5153f56bcfa req-236c5f89-e968-4acd-a363-aa41cf291cf3 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Updated VIF entry in instance network info cache for port ed08f6b9-7027-417b-91d5-e9639b3fed92. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1016.637643] env[61243]: DEBUG nova.network.neutron [req-ee07792f-7845-4425-83ec-e5153f56bcfa req-236c5f89-e968-4acd-a363-aa41cf291cf3 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Updating instance_info_cache with network_info: [{"id": "ed08f6b9-7027-417b-91d5-e9639b3fed92", "address": "fa:16:3e:8f:9c:75", "network": {"id": "523d4cf1-467c-4ea6-80a4-ebbd01016efa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-919325376-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7651b565d1647eb948fe07faba32a41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped08f6b9-70", "ovs_interfaceid": "ed08f6b9-7027-417b-91d5-e9639b3fed92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.854498] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.797s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.857323] env[61243]: DEBUG oslo_concurrency.lockutils [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.565s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.857451] env[61243]: DEBUG nova.objects.instance [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lazy-loading 'resources' on Instance uuid 7ca74bdc-c272-4b6e-a4dd-56d97144b197 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.878393] env[61243]: INFO nova.scheduler.client.report [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Deleted allocations for instance b0b5b215-d7ff-4960-9d39-cf792d8d4656 [ 1016.900930] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': task-1339349, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077103} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.901267] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1016.902088] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-803ff945-dad8-48a0-86e2-e575416cbce7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.924761] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 0e03eee3-e7bc-483f-948f-a5b408375e24/0e03eee3-e7bc-483f-948f-a5b408375e24.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1016.925378] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2f76930-42b1-418f-adf9-2b271e1a9246 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.940195] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "refresh_cache-2c4217cd-af4a-4d64-a422-73955326df03" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.946872] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Waiting for the task: (returnval){ [ 1016.946872] env[61243]: value = "task-1339350" [ 1016.946872] env[61243]: _type = "Task" [ 1016.946872] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.955995] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': task-1339350, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.026107] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.026465] env[61243]: DEBUG nova.compute.manager [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Instance network_info: |[{"id": "f2a48425-d436-40c9-a2dc-694e6313da6d", "address": "fa:16:3e:34:84:e2", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2a48425-d4", "ovs_interfaceid": "f2a48425-d436-40c9-a2dc-694e6313da6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1017.026752] env[61243]: DEBUG oslo_concurrency.lockutils [req-14312bca-422d-40c4-ad8f-86da87da2fe4 req-be0b42bf-752a-4802-8139-7dce513b7cd7 service nova] Acquired lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.026944] env[61243]: DEBUG nova.network.neutron [req-14312bca-422d-40c4-ad8f-86da87da2fe4 req-be0b42bf-752a-4802-8139-7dce513b7cd7 service nova] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Refreshing network info cache for port f2a48425-d436-40c9-a2dc-694e6313da6d {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1017.028196] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:34:84:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd7b5f1ef-d4b9-4ec3-b047-17e4cb349d25', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f2a48425-d436-40c9-a2dc-694e6313da6d', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1017.035961] env[61243]: DEBUG oslo.service.loopingcall [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.036904] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1017.037195] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-53256cf8-c291-4afd-a2c4-8d7717ff843f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.056705] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1017.056705] env[61243]: value = "task-1339351" [ 1017.056705] env[61243]: _type = "Task" [ 1017.056705] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.064729] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339351, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.140630] env[61243]: DEBUG oslo_concurrency.lockutils [req-ee07792f-7845-4425-83ec-e5153f56bcfa req-236c5f89-e968-4acd-a363-aa41cf291cf3 service nova] Releasing lock "refresh_cache-dca3301d-ef3e-48ef-920c-866b4a086ea6" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.392858] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c03e1f7c-2b0b-4691-bfb1-21037dc7485c tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "b0b5b215-d7ff-4960-9d39-cf792d8d4656" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.173s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.464543] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': task-1339350, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.540172] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242e2d43-c50b-4cef-8bd9-8de36ab7ba59 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.547803] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c6c15d-374a-4e25-958c-143ae6b406d0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.587424] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90118b82-df64-4c99-a84b-ff6bd75202f4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.598198] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cde19dd3-21ff-4463-9f31-847a1b4b13c8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.602015] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339351, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.612238] env[61243]: DEBUG nova.compute.provider_tree [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1017.758428] env[61243]: DEBUG nova.network.neutron [req-14312bca-422d-40c4-ad8f-86da87da2fe4 req-be0b42bf-752a-4802-8139-7dce513b7cd7 service nova] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updated VIF entry in instance network info cache for port f2a48425-d436-40c9-a2dc-694e6313da6d. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1017.758821] env[61243]: DEBUG nova.network.neutron [req-14312bca-422d-40c4-ad8f-86da87da2fe4 req-be0b42bf-752a-4802-8139-7dce513b7cd7 service nova] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating instance_info_cache with network_info: [{"id": "f2a48425-d436-40c9-a2dc-694e6313da6d", "address": "fa:16:3e:34:84:e2", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2a48425-d4", "ovs_interfaceid": "f2a48425-d436-40c9-a2dc-694e6313da6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.956970] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "5660de64-7152-4832-960e-2deb1c0bc37e" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.957786] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "5660de64-7152-4832-960e-2deb1c0bc37e" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.958142] env[61243]: DEBUG nova.compute.manager [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Going to confirm migration 2 {{(pid=61243) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1017.963992] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': task-1339350, 'name': ReconfigVM_Task, 'duration_secs': 0.622534} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.964285] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 0e03eee3-e7bc-483f-948f-a5b408375e24/0e03eee3-e7bc-483f-948f-a5b408375e24.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1017.965113] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-23f1f0bf-c73e-41eb-b43c-55e07b91be3d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.972698] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Waiting for the task: (returnval){ [ 1017.972698] env[61243]: value = "task-1339352" [ 1017.972698] env[61243]: _type = "Task" [ 1017.972698] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.980426] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': task-1339352, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.091905] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339351, 'name': CreateVM_Task, 'duration_secs': 0.548678} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.092151] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1018.093116] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.093337] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.093874] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1018.094213] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c4f0a9b-cbca-4524-bcb6-0db08755c85f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.099364] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1018.099364] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e5c305-0db5-aa6c-d7eb-a3f1fbe4e48f" [ 1018.099364] env[61243]: _type = "Task" [ 1018.099364] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.107374] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e5c305-0db5-aa6c-d7eb-a3f1fbe4e48f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.144980] env[61243]: DEBUG nova.scheduler.client.report [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 136 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1018.145308] env[61243]: DEBUG nova.compute.provider_tree [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 136 to 137 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1018.145527] env[61243]: DEBUG nova.compute.provider_tree [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1018.261225] env[61243]: DEBUG oslo_concurrency.lockutils [req-14312bca-422d-40c4-ad8f-86da87da2fe4 req-be0b42bf-752a-4802-8139-7dce513b7cd7 service nova] Releasing lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.471722] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762eddcd-4da6-4702-b242-4c75077eae18 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.489266] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': task-1339352, 'name': Rename_Task, 'duration_secs': 0.3397} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.502062] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1018.502440] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Updating instance '2c4217cd-af4a-4d64-a422-73955326df03' progress to 0 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1018.505860] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-02bad2ae-83f1-446d-bef5-7fd9197e401e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.514256] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Waiting for the task: (returnval){ [ 1018.514256] env[61243]: value = "task-1339353" [ 1018.514256] env[61243]: _type = "Task" [ 1018.514256] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.522318] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': task-1339353, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.547703] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "refresh_cache-5660de64-7152-4832-960e-2deb1c0bc37e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.547928] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "refresh_cache-5660de64-7152-4832-960e-2deb1c0bc37e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.548159] env[61243]: DEBUG nova.network.neutron [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1018.548357] env[61243]: DEBUG nova.objects.instance [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lazy-loading 'info_cache' on Instance uuid 5660de64-7152-4832-960e-2deb1c0bc37e {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.610688] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e5c305-0db5-aa6c-d7eb-a3f1fbe4e48f, 'name': SearchDatastore_Task, 'duration_secs': 0.026711} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.611201] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.611460] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.611716] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.611875] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.612075] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.612377] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fff9182e-3bda-4e63-8304-d262f8d86def {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.621870] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.622067] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1018.622761] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a3193ed-5037-4c92-8246-1a4a6885e696 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.627869] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1018.627869] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522baaa0-0288-2288-9254-4d90d599b0b7" [ 1018.627869] env[61243]: _type = "Task" [ 1018.627869] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.635394] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522baaa0-0288-2288-9254-4d90d599b0b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.650207] env[61243]: DEBUG oslo_concurrency.lockutils [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.793s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.652335] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.252s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.652578] env[61243]: DEBUG nova.objects.instance [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lazy-loading 'resources' on Instance uuid c1a0d242-4eae-4c03-8341-840b41341f17 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.674385] env[61243]: INFO nova.scheduler.client.report [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleted allocations for instance 7ca74bdc-c272-4b6e-a4dd-56d97144b197 [ 1019.012262] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1019.012561] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8950f4b6-e070-492f-966b-c847368f3642 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.022990] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': task-1339353, 'name': PowerOnVM_Task} progress is 71%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.024267] env[61243]: DEBUG oslo_vmware.api [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 1019.024267] env[61243]: value = "task-1339354" [ 1019.024267] env[61243]: _type = "Task" [ 1019.024267] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.031529] env[61243]: DEBUG oslo_vmware.api [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339354, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.138098] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]522baaa0-0288-2288-9254-4d90d599b0b7, 'name': SearchDatastore_Task, 'duration_secs': 0.01877} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.138887] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7a88b3a-9a94-4516-84b1-ed6dded774c1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.144082] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1019.144082] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c8d788-f1cd-66cd-d18a-7d669d48f9d9" [ 1019.144082] env[61243]: _type = "Task" [ 1019.144082] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.152965] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c8d788-f1cd-66cd-d18a-7d669d48f9d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.183842] env[61243]: DEBUG oslo_concurrency.lockutils [None req-36fb8c2a-4e67-485c-9265-ffbb805eeb8a tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "7ca74bdc-c272-4b6e-a4dd-56d97144b197" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.815s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.313772] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2a58e06-60be-4afa-b3a5-fa91fb1ff32c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.320966] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94de1c88-a37a-4176-968f-d83eadb1b32c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.350895] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215037be-db25-4f67-873a-e83c03cb7676 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.358092] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f2ac43-97ef-454f-9871-0ac2043c0ae0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.371675] env[61243]: DEBUG nova.compute.provider_tree [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.524483] env[61243]: DEBUG oslo_vmware.api [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': task-1339353, 'name': PowerOnVM_Task, 'duration_secs': 0.774587} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.525066] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1019.525310] env[61243]: INFO nova.compute.manager [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Took 7.05 seconds to spawn the instance on the hypervisor. [ 1019.525501] env[61243]: DEBUG nova.compute.manager [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1019.528892] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21a5f51-a9f8-445f-9be5-946fba5f9fb2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.535757] env[61243]: DEBUG oslo_vmware.api [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339354, 'name': PowerOffVM_Task, 'duration_secs': 0.180517} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.537246] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1019.537449] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Updating instance '2c4217cd-af4a-4d64-a422-73955326df03' progress to 17 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1019.654379] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c8d788-f1cd-66cd-d18a-7d669d48f9d9, 'name': SearchDatastore_Task, 'duration_secs': 0.010394} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.656811] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.657109] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 60fa2863-f0aa-4798-8e5f-e93df47f93e8/60fa2863-f0aa-4798-8e5f-e93df47f93e8.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1019.657389] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3db0bbf8-1093-45c0-9c87-8ba9649e03eb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.664354] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1019.664354] env[61243]: value = "task-1339355" [ 1019.664354] env[61243]: _type = "Task" [ 1019.664354] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.673685] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339355, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.754153] env[61243]: DEBUG nova.network.neutron [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Updating instance_info_cache with network_info: [{"id": "2b065192-815c-492f-8f79-bfa9553ff080", "address": "fa:16:3e:94:3e:9e", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b065192-81", "ovs_interfaceid": "2b065192-815c-492f-8f79-bfa9553ff080", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.874695] env[61243]: DEBUG nova.scheduler.client.report [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1020.010623] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.010869] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.043597] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1020.043990] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1020.044122] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1020.044321] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1020.044484] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1020.044628] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1020.044984] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1020.045254] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1020.045454] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1020.045629] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1020.045814] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1020.056188] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21f563f5-02de-4190-afd1-7c7f99964cfa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.073722] env[61243]: INFO nova.compute.manager [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Took 14.00 seconds to build instance. [ 1020.075576] env[61243]: DEBUG oslo_concurrency.lockutils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.075843] env[61243]: DEBUG oslo_concurrency.lockutils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.081288] env[61243]: DEBUG oslo_vmware.api [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 1020.081288] env[61243]: value = "task-1339356" [ 1020.081288] env[61243]: _type = "Task" [ 1020.081288] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.093655] env[61243]: DEBUG oslo_vmware.api [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339356, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.177063] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339355, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.256942] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "refresh_cache-5660de64-7152-4832-960e-2deb1c0bc37e" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.257255] env[61243]: DEBUG nova.objects.instance [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lazy-loading 'migration_context' on Instance uuid 5660de64-7152-4832-960e-2deb1c0bc37e {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.379969] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.727s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.401661] env[61243]: INFO nova.scheduler.client.report [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Deleted allocations for instance c1a0d242-4eae-4c03-8341-840b41341f17 [ 1020.514364] env[61243]: DEBUG nova.compute.manager [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1020.578566] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0b08afa2-68b8-42da-8665-b89a2673fd6d tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Lock "0e03eee3-e7bc-483f-948f-a5b408375e24" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.518s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.582139] env[61243]: DEBUG nova.compute.manager [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1020.594022] env[61243]: DEBUG oslo_vmware.api [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339356, 'name': ReconfigVM_Task, 'duration_secs': 0.499813} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.594461] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Updating instance '2c4217cd-af4a-4d64-a422-73955326df03' progress to 33 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1020.677156] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339355, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.570377} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.677376] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 60fa2863-f0aa-4798-8e5f-e93df47f93e8/60fa2863-f0aa-4798-8e5f-e93df47f93e8.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1020.677627] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1020.677846] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-82a74dbf-e9ad-40b7-a4d5-1079a6f7c8a0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.684445] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1020.684445] env[61243]: value = "task-1339357" [ 1020.684445] env[61243]: _type = "Task" [ 1020.684445] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.692142] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339357, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.761197] env[61243]: DEBUG nova.objects.base [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Object Instance<5660de64-7152-4832-960e-2deb1c0bc37e> lazy-loaded attributes: info_cache,migration_context {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1020.762156] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27253d44-e2d0-45b2-ab2d-20bc362b37a3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.781910] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-438d861a-ead4-451d-86ec-bec4d318a018 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.787567] env[61243]: DEBUG oslo_vmware.api [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1020.787567] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529edd4a-39ba-bfea-ce1e-4d953760b201" [ 1020.787567] env[61243]: _type = "Task" [ 1020.787567] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.795492] env[61243]: DEBUG oslo_vmware.api [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529edd4a-39ba-bfea-ce1e-4d953760b201, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.911987] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c8cd4a67-d366-4f9f-afd5-e4d299f15966 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "c1a0d242-4eae-4c03-8341-840b41341f17" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.279s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.035062] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.035062] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.036382] env[61243]: INFO nova.compute.claims [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1021.100061] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1021.100502] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1021.100502] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1021.100624] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1021.100785] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1021.101022] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1021.101280] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1021.101454] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1021.101671] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1021.101875] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1021.102616] env[61243]: DEBUG nova.virt.hardware [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1021.108194] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Reconfiguring VM instance instance-0000005f to detach disk 2000 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1021.109480] env[61243]: DEBUG oslo_concurrency.lockutils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.110259] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68893309-ddf4-47ab-b973-0aee1e7283f7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.123734] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Acquiring lock "0e03eee3-e7bc-483f-948f-a5b408375e24" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.124010] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Lock "0e03eee3-e7bc-483f-948f-a5b408375e24" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.124246] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Acquiring lock "0e03eee3-e7bc-483f-948f-a5b408375e24-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.124443] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Lock "0e03eee3-e7bc-483f-948f-a5b408375e24-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.124621] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Lock "0e03eee3-e7bc-483f-948f-a5b408375e24-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.126762] env[61243]: INFO nova.compute.manager [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Terminating instance [ 1021.128414] env[61243]: DEBUG nova.compute.manager [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1021.128613] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1021.129386] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4efeba4-60b8-4e21-9607-8eb56e3368dd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.133117] env[61243]: DEBUG oslo_vmware.api [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 1021.133117] env[61243]: value = "task-1339358" [ 1021.133117] env[61243]: _type = "Task" [ 1021.133117] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.142691] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1021.143179] env[61243]: DEBUG oslo_vmware.api [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339358, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.143389] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4260c57d-702e-4d33-bf00-bb458c0271e0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.148545] env[61243]: DEBUG oslo_vmware.api [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Waiting for the task: (returnval){ [ 1021.148545] env[61243]: value = "task-1339359" [ 1021.148545] env[61243]: _type = "Task" [ 1021.148545] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.157763] env[61243]: DEBUG oslo_vmware.api [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': task-1339359, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.194740] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339357, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070227} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.195081] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1021.195895] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0553c88-abbc-472a-96e6-de20e60efd4d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.219782] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 60fa2863-f0aa-4798-8e5f-e93df47f93e8/60fa2863-f0aa-4798-8e5f-e93df47f93e8.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.220152] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e366fdc-2a0a-4df8-b71f-543347fc9f2c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.240917] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1021.240917] env[61243]: value = "task-1339360" [ 1021.240917] env[61243]: _type = "Task" [ 1021.240917] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.252231] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339360, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.297850] env[61243]: DEBUG oslo_vmware.api [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529edd4a-39ba-bfea-ce1e-4d953760b201, 'name': SearchDatastore_Task, 'duration_secs': 0.041895} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.298212] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.643499] env[61243]: DEBUG oslo_vmware.api [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339358, 'name': ReconfigVM_Task, 'duration_secs': 0.234294} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.643796] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Reconfigured VM instance instance-0000005f to detach disk 2000 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1021.644657] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab235f9-fc9d-45eb-a8dc-d27c80d117cc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.666501] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 2c4217cd-af4a-4d64-a422-73955326df03/2c4217cd-af4a-4d64-a422-73955326df03.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.669630] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c8c973d-e634-4877-b316-97b2704ebb8a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.687801] env[61243]: DEBUG oslo_vmware.api [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': task-1339359, 'name': PowerOffVM_Task, 'duration_secs': 0.192881} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.689085] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1021.689284] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1021.689596] env[61243]: DEBUG oslo_vmware.api [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 1021.689596] env[61243]: value = "task-1339361" [ 1021.689596] env[61243]: _type = "Task" [ 1021.689596] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.689795] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1cbb2ee0-54ec-42ce-866c-fd24fc174d24 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.699024] env[61243]: DEBUG oslo_vmware.api [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339361, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.749966] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339360, 'name': ReconfigVM_Task, 'duration_secs': 0.349734} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.750384] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 60fa2863-f0aa-4798-8e5f-e93df47f93e8/60fa2863-f0aa-4798-8e5f-e93df47f93e8.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.751078] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-21adaf7b-97d1-4e2c-b010-c0b3f4c6da97 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.757495] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1021.757495] env[61243]: value = "task-1339363" [ 1021.757495] env[61243]: _type = "Task" [ 1021.757495] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.765527] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339363, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.801117] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1021.801197] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1021.801520] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Deleting the datastore file [datastore2] 0e03eee3-e7bc-483f-948f-a5b408375e24 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1021.801930] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ac270367-c7fb-4bc6-9ee9-f0dcdfd8dd79 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.808626] env[61243]: DEBUG oslo_vmware.api [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Waiting for the task: (returnval){ [ 1021.808626] env[61243]: value = "task-1339364" [ 1021.808626] env[61243]: _type = "Task" [ 1021.808626] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.818242] env[61243]: DEBUG oslo_vmware.api [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': task-1339364, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.200094] env[61243]: DEBUG oslo_vmware.api [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339361, 'name': ReconfigVM_Task, 'duration_secs': 0.356775} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.201318] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 2c4217cd-af4a-4d64-a422-73955326df03/2c4217cd-af4a-4d64-a422-73955326df03.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1022.201420] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Updating instance '2c4217cd-af4a-4d64-a422-73955326df03' progress to 50 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1022.205246] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe56cfb-10ba-4afa-8876-a38170e7100d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.212324] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe2343f-6fbc-4799-8e0c-42e49273ebc3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.241931] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb483afb-e5f4-4a8e-b68a-433bc10698fc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.249682] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8ba06c-3795-4642-8bc6-632b830a2712 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.263349] env[61243]: DEBUG nova.compute.provider_tree [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1022.272791] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339363, 'name': Rename_Task, 'duration_secs': 0.192732} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.274066] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1022.274066] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1f7be40-65bf-4736-8973-73225118e44c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.280402] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1022.280402] env[61243]: value = "task-1339365" [ 1022.280402] env[61243]: _type = "Task" [ 1022.280402] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.288112] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339365, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.319100] env[61243]: DEBUG oslo_vmware.api [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Task: {'id': task-1339364, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.232824} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.319440] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1022.319604] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1022.319886] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1022.320009] env[61243]: INFO nova.compute.manager [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1022.320285] env[61243]: DEBUG oslo.service.loopingcall [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.320498] env[61243]: DEBUG nova.compute.manager [-] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1022.320608] env[61243]: DEBUG nova.network.neutron [-] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1022.588012] env[61243]: DEBUG nova.compute.manager [req-5dba31e1-249f-4a46-8041-5d89c4ddbfcf req-131860ca-93d9-4886-b8c3-b88497b62bf3 service nova] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Received event network-vif-deleted-b3d60ea2-8847-41fc-aa33-1998bb27a115 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1022.588378] env[61243]: INFO nova.compute.manager [req-5dba31e1-249f-4a46-8041-5d89c4ddbfcf req-131860ca-93d9-4886-b8c3-b88497b62bf3 service nova] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Neutron deleted interface b3d60ea2-8847-41fc-aa33-1998bb27a115; detaching it from the instance and deleting it from the info cache [ 1022.588553] env[61243]: DEBUG nova.network.neutron [req-5dba31e1-249f-4a46-8041-5d89c4ddbfcf req-131860ca-93d9-4886-b8c3-b88497b62bf3 service nova] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.710272] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0b459d-6449-4d64-8681-a596fe5ff02b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.730255] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f77aebb-e2c1-4bba-828b-dba711ff6070 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.750327] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Updating instance '2c4217cd-af4a-4d64-a422-73955326df03' progress to 67 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1022.787032] env[61243]: ERROR nova.scheduler.client.report [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [req-ed3f829e-c0c7-4dc6-ada0-2a5b37cf0af7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ed3f829e-c0c7-4dc6-ada0-2a5b37cf0af7"}]} [ 1022.792490] env[61243]: DEBUG oslo_vmware.api [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339365, 'name': PowerOnVM_Task, 'duration_secs': 0.479997} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.793471] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1022.793688] env[61243]: INFO nova.compute.manager [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Took 7.97 seconds to spawn the instance on the hypervisor. [ 1022.793874] env[61243]: DEBUG nova.compute.manager [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.794687] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad11d2d0-a904-440e-8a59-c9964d673d9d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.804410] env[61243]: DEBUG nova.scheduler.client.report [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1022.816426] env[61243]: DEBUG nova.scheduler.client.report [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1022.816673] env[61243]: DEBUG nova.compute.provider_tree [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1022.827663] env[61243]: DEBUG nova.scheduler.client.report [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1022.844196] env[61243]: DEBUG nova.scheduler.client.report [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1023.001782] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3acb420-5dc6-48d4-b65a-88caaffb4cb9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.009205] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67861bdc-875d-4a08-9dbc-c78d5080d257 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.037971] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-996506c7-bcd9-4526-9989-c9f73e213287 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.046291] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab8d839-1ed4-4db9-8a0b-d5bedb3ebb82 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.058366] env[61243]: DEBUG nova.compute.provider_tree [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1023.059751] env[61243]: DEBUG nova.network.neutron [-] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.091333] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-419181d7-12f3-4abb-9495-f5eb79b44b60 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.100730] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c95aae-e7c2-4d2f-998f-a2c3d6bb5182 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.130416] env[61243]: DEBUG nova.compute.manager [req-5dba31e1-249f-4a46-8041-5d89c4ddbfcf req-131860ca-93d9-4886-b8c3-b88497b62bf3 service nova] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Detach interface failed, port_id=b3d60ea2-8847-41fc-aa33-1998bb27a115, reason: Instance 0e03eee3-e7bc-483f-948f-a5b408375e24 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1023.267484] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "486ebf65-4376-424c-9deb-f1ef70e44be3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.267484] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.311734] env[61243]: INFO nova.compute.manager [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Took 16.31 seconds to build instance. [ 1023.316512] env[61243]: DEBUG nova.network.neutron [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Port 611257cf-50e1-491a-beae-8c942398013d binding to destination host cpu-1 is already ACTIVE {{(pid=61243) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1023.565641] env[61243]: INFO nova.compute.manager [-] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Took 1.24 seconds to deallocate network for instance. [ 1023.599788] env[61243]: DEBUG nova.scheduler.client.report [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 140 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1023.600075] env[61243]: DEBUG nova.compute.provider_tree [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 140 to 141 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1023.600272] env[61243]: DEBUG nova.compute.provider_tree [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1023.769821] env[61243]: DEBUG nova.compute.manager [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1023.812305] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f4f761a7-44f1-4f18-8d83-2f1946f405ea tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.847s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.074874] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.106044] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.071s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.106604] env[61243]: DEBUG nova.compute.manager [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1024.109872] env[61243]: DEBUG oslo_concurrency.lockutils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.111532] env[61243]: INFO nova.compute.claims [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1024.291593] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.338139] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "2c4217cd-af4a-4d64-a422-73955326df03-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.338375] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "2c4217cd-af4a-4d64-a422-73955326df03-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.338558] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "2c4217cd-af4a-4d64-a422-73955326df03-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.611430] env[61243]: DEBUG nova.compute.utils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1024.614044] env[61243]: DEBUG nova.compute.manager [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1024.616034] env[61243]: DEBUG nova.network.neutron [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1024.618495] env[61243]: DEBUG nova.compute.manager [req-697b2025-a8e0-4d6a-8666-4559a0df4ee2 req-7241db5b-a638-4adb-9953-fab8ba1b8854 service nova] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Received event network-changed-f2a48425-d436-40c9-a2dc-694e6313da6d {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1024.618685] env[61243]: DEBUG nova.compute.manager [req-697b2025-a8e0-4d6a-8666-4559a0df4ee2 req-7241db5b-a638-4adb-9953-fab8ba1b8854 service nova] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Refreshing instance network info cache due to event network-changed-f2a48425-d436-40c9-a2dc-694e6313da6d. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1024.619032] env[61243]: DEBUG oslo_concurrency.lockutils [req-697b2025-a8e0-4d6a-8666-4559a0df4ee2 req-7241db5b-a638-4adb-9953-fab8ba1b8854 service nova] Acquiring lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.619105] env[61243]: DEBUG oslo_concurrency.lockutils [req-697b2025-a8e0-4d6a-8666-4559a0df4ee2 req-7241db5b-a638-4adb-9953-fab8ba1b8854 service nova] Acquired lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.619315] env[61243]: DEBUG nova.network.neutron [req-697b2025-a8e0-4d6a-8666-4559a0df4ee2 req-7241db5b-a638-4adb-9953-fab8ba1b8854 service nova] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Refreshing network info cache for port f2a48425-d436-40c9-a2dc-694e6313da6d {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1024.666270] env[61243]: DEBUG nova.policy [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b32f4788825f4a8e82e028cc10347746', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '683c027e4adc4cc4a6b8747a8c8c0de9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1024.927688] env[61243]: DEBUG nova.network.neutron [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Successfully created port: de6bbdd5-c1ee-40d7-a91c-1938f53863a7 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1025.121083] env[61243]: DEBUG nova.compute.manager [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1025.309361] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5842d06-96fb-4704-94bd-e791f5b988e1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.318195] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b85b71-5603-4a5b-9c71-e7a136c216ca {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.353149] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c608dd52-1ff1-4717-adc6-44c7611d1f6b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.361057] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4687994-d015-4f22-8f98-5608bad47c37 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.375164] env[61243]: DEBUG nova.compute.provider_tree [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.398994] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "refresh_cache-2c4217cd-af4a-4d64-a422-73955326df03" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.399214] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "refresh_cache-2c4217cd-af4a-4d64-a422-73955326df03" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.399406] env[61243]: DEBUG nova.network.neutron [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1025.414030] env[61243]: DEBUG nova.network.neutron [req-697b2025-a8e0-4d6a-8666-4559a0df4ee2 req-7241db5b-a638-4adb-9953-fab8ba1b8854 service nova] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updated VIF entry in instance network info cache for port f2a48425-d436-40c9-a2dc-694e6313da6d. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1025.414497] env[61243]: DEBUG nova.network.neutron [req-697b2025-a8e0-4d6a-8666-4559a0df4ee2 req-7241db5b-a638-4adb-9953-fab8ba1b8854 service nova] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating instance_info_cache with network_info: [{"id": "f2a48425-d436-40c9-a2dc-694e6313da6d", "address": "fa:16:3e:34:84:e2", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2a48425-d4", "ovs_interfaceid": "f2a48425-d436-40c9-a2dc-694e6313da6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.878818] env[61243]: DEBUG nova.scheduler.client.report [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1025.916573] env[61243]: DEBUG oslo_concurrency.lockutils [req-697b2025-a8e0-4d6a-8666-4559a0df4ee2 req-7241db5b-a638-4adb-9953-fab8ba1b8854 service nova] Releasing lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.092075] env[61243]: DEBUG nova.network.neutron [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Updating instance_info_cache with network_info: [{"id": "611257cf-50e1-491a-beae-8c942398013d", "address": "fa:16:3e:6a:c8:a1", "network": {"id": "0788f9ac-5a15-4431-a711-9b4f6cf79fbc", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1210957657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "113713170fb94b8a8d7985b9a02faf77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c7821ea-f92f-4f06-a4cb-05e1186a9d22", "external-id": "nsx-vlan-transportzone-69", "segmentation_id": 69, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap611257cf-50", "ovs_interfaceid": "611257cf-50e1-491a-beae-8c942398013d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.135893] env[61243]: DEBUG nova.compute.manager [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1026.161420] env[61243]: DEBUG nova.virt.hardware [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1026.161664] env[61243]: DEBUG nova.virt.hardware [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1026.161885] env[61243]: DEBUG nova.virt.hardware [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1026.162128] env[61243]: DEBUG nova.virt.hardware [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1026.162291] env[61243]: DEBUG nova.virt.hardware [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1026.162447] env[61243]: DEBUG nova.virt.hardware [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1026.162660] env[61243]: DEBUG nova.virt.hardware [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1026.162824] env[61243]: DEBUG nova.virt.hardware [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1026.162995] env[61243]: DEBUG nova.virt.hardware [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1026.163183] env[61243]: DEBUG nova.virt.hardware [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1026.163365] env[61243]: DEBUG nova.virt.hardware [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1026.164279] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b699f5f0-c950-4fcd-a03e-7c7b0fef2a4a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.172619] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e6d169c-94e9-4876-97ff-d1ad9afacb5e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.330889] env[61243]: DEBUG nova.compute.manager [req-a8227e1a-71b0-4f95-bce0-915d62744609 req-d794cc60-087f-47a2-be27-0f4f257ca083 service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Received event network-vif-plugged-de6bbdd5-c1ee-40d7-a91c-1938f53863a7 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.331160] env[61243]: DEBUG oslo_concurrency.lockutils [req-a8227e1a-71b0-4f95-bce0-915d62744609 req-d794cc60-087f-47a2-be27-0f4f257ca083 service nova] Acquiring lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.331362] env[61243]: DEBUG oslo_concurrency.lockutils [req-a8227e1a-71b0-4f95-bce0-915d62744609 req-d794cc60-087f-47a2-be27-0f4f257ca083 service nova] Lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.331566] env[61243]: DEBUG oslo_concurrency.lockutils [req-a8227e1a-71b0-4f95-bce0-915d62744609 req-d794cc60-087f-47a2-be27-0f4f257ca083 service nova] Lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.331782] env[61243]: DEBUG nova.compute.manager [req-a8227e1a-71b0-4f95-bce0-915d62744609 req-d794cc60-087f-47a2-be27-0f4f257ca083 service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] No waiting events found dispatching network-vif-plugged-de6bbdd5-c1ee-40d7-a91c-1938f53863a7 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1026.331972] env[61243]: WARNING nova.compute.manager [req-a8227e1a-71b0-4f95-bce0-915d62744609 req-d794cc60-087f-47a2-be27-0f4f257ca083 service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Received unexpected event network-vif-plugged-de6bbdd5-c1ee-40d7-a91c-1938f53863a7 for instance with vm_state building and task_state spawning. [ 1026.383826] env[61243]: DEBUG oslo_concurrency.lockutils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.274s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.384585] env[61243]: DEBUG nova.compute.manager [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1026.387238] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 5.089s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.595074] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "refresh_cache-2c4217cd-af4a-4d64-a422-73955326df03" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.892130] env[61243]: DEBUG nova.compute.utils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1026.896646] env[61243]: DEBUG nova.network.neutron [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Successfully updated port: de6bbdd5-c1ee-40d7-a91c-1938f53863a7 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1026.900111] env[61243]: DEBUG nova.compute.manager [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1026.900111] env[61243]: DEBUG nova.network.neutron [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1026.940017] env[61243]: DEBUG nova.policy [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06e247d66c8d43218302a35ead439405', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e641e72e4d2048218ef403456dbb2b82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1026.958232] env[61243]: DEBUG nova.compute.manager [req-4c7171d6-f3e8-466e-b1bb-aa3972d8b95a req-b1ab3a55-d882-43d2-b5c7-8354b534a7d8 service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Received event network-changed-de6bbdd5-c1ee-40d7-a91c-1938f53863a7 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.958448] env[61243]: DEBUG nova.compute.manager [req-4c7171d6-f3e8-466e-b1bb-aa3972d8b95a req-b1ab3a55-d882-43d2-b5c7-8354b534a7d8 service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Refreshing instance network info cache due to event network-changed-de6bbdd5-c1ee-40d7-a91c-1938f53863a7. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1026.958666] env[61243]: DEBUG oslo_concurrency.lockutils [req-4c7171d6-f3e8-466e-b1bb-aa3972d8b95a req-b1ab3a55-d882-43d2-b5c7-8354b534a7d8 service nova] Acquiring lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.959324] env[61243]: DEBUG oslo_concurrency.lockutils [req-4c7171d6-f3e8-466e-b1bb-aa3972d8b95a req-b1ab3a55-d882-43d2-b5c7-8354b534a7d8 service nova] Acquired lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.959413] env[61243]: DEBUG nova.network.neutron [req-4c7171d6-f3e8-466e-b1bb-aa3972d8b95a req-b1ab3a55-d882-43d2-b5c7-8354b534a7d8 service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Refreshing network info cache for port de6bbdd5-c1ee-40d7-a91c-1938f53863a7 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1027.063065] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1276a8-9357-486f-b8b2-00e15488917c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.070291] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53cc556d-ca94-4cc1-869f-87ce54b694ef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.104137] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-398f08a7-0ab3-4cd7-bd58-622a89b7a1ac {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.113834] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab6116a-0268-43ff-b9f7-fcffbb98809f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.119407] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8abd12b-b0ec-461f-a02d-876cde48f9e9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.129758] env[61243]: DEBUG nova.compute.provider_tree [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.146627] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b212f627-daf3-47ae-ad4e-5d11f5ba6e70 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.153734] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Updating instance '2c4217cd-af4a-4d64-a422-73955326df03' progress to 83 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1027.197304] env[61243]: DEBUG nova.network.neutron [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Successfully created port: 58c44c97-32fa-4f2f-9313-6fee30a10aec {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1027.397105] env[61243]: DEBUG nova.compute.manager [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1027.400332] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.489787] env[61243]: DEBUG nova.network.neutron [req-4c7171d6-f3e8-466e-b1bb-aa3972d8b95a req-b1ab3a55-d882-43d2-b5c7-8354b534a7d8 service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1027.561402] env[61243]: DEBUG nova.network.neutron [req-4c7171d6-f3e8-466e-b1bb-aa3972d8b95a req-b1ab3a55-d882-43d2-b5c7-8354b534a7d8 service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.633536] env[61243]: DEBUG nova.scheduler.client.report [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1027.660161] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1027.660486] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e862ba0-97e3-4263-85cf-65ec54122d24 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.668775] env[61243]: DEBUG oslo_vmware.api [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 1027.668775] env[61243]: value = "task-1339366" [ 1027.668775] env[61243]: _type = "Task" [ 1027.668775] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.676383] env[61243]: DEBUG oslo_vmware.api [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339366, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.065078] env[61243]: DEBUG oslo_concurrency.lockutils [req-4c7171d6-f3e8-466e-b1bb-aa3972d8b95a req-b1ab3a55-d882-43d2-b5c7-8354b534a7d8 service nova] Releasing lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.065368] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.065568] env[61243]: DEBUG nova.network.neutron [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1028.178419] env[61243]: DEBUG oslo_vmware.api [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339366, 'name': PowerOnVM_Task, 'duration_secs': 0.423059} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.178771] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1028.178878] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6e751d28-a5fd-46b4-8f06-b50ccf6845bc tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Updating instance '2c4217cd-af4a-4d64-a422-73955326df03' progress to 100 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1028.410981] env[61243]: DEBUG nova.compute.manager [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1028.438915] env[61243]: DEBUG nova.virt.hardware [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1028.439190] env[61243]: DEBUG nova.virt.hardware [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1028.439358] env[61243]: DEBUG nova.virt.hardware [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1028.439553] env[61243]: DEBUG nova.virt.hardware [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1028.439707] env[61243]: DEBUG nova.virt.hardware [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1028.439859] env[61243]: DEBUG nova.virt.hardware [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1028.440082] env[61243]: DEBUG nova.virt.hardware [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1028.440334] env[61243]: DEBUG nova.virt.hardware [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1028.440608] env[61243]: DEBUG nova.virt.hardware [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1028.440691] env[61243]: DEBUG nova.virt.hardware [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1028.440834] env[61243]: DEBUG nova.virt.hardware [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1028.441771] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa1b9a0-220b-45b5-9498-970a7e874126 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.449464] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de2b2cfd-ff77-4714-9df9-544e3b6b5040 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.613297] env[61243]: DEBUG nova.network.neutron [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1028.643963] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.257s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.646939] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.572s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.647279] env[61243]: DEBUG nova.objects.instance [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Lazy-loading 'resources' on Instance uuid 0e03eee3-e7bc-483f-948f-a5b408375e24 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.759251] env[61243]: DEBUG nova.network.neutron [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Successfully updated port: 58c44c97-32fa-4f2f-9313-6fee30a10aec {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1028.806789] env[61243]: DEBUG nova.network.neutron [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Updating instance_info_cache with network_info: [{"id": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "address": "fa:16:3e:43:9a:6f", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde6bbdd5-c1", "ovs_interfaceid": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.984749] env[61243]: DEBUG nova.compute.manager [req-3157c1cb-7ce9-4ae2-ab7e-ac4c2aa6aad1 req-ac3632d3-e712-4243-bc3b-2658696a1bae service nova] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Received event network-vif-plugged-58c44c97-32fa-4f2f-9313-6fee30a10aec {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.984989] env[61243]: DEBUG oslo_concurrency.lockutils [req-3157c1cb-7ce9-4ae2-ab7e-ac4c2aa6aad1 req-ac3632d3-e712-4243-bc3b-2658696a1bae service nova] Acquiring lock "05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.985733] env[61243]: DEBUG oslo_concurrency.lockutils [req-3157c1cb-7ce9-4ae2-ab7e-ac4c2aa6aad1 req-ac3632d3-e712-4243-bc3b-2658696a1bae service nova] Lock "05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.985957] env[61243]: DEBUG oslo_concurrency.lockutils [req-3157c1cb-7ce9-4ae2-ab7e-ac4c2aa6aad1 req-ac3632d3-e712-4243-bc3b-2658696a1bae service nova] Lock "05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.986183] env[61243]: DEBUG nova.compute.manager [req-3157c1cb-7ce9-4ae2-ab7e-ac4c2aa6aad1 req-ac3632d3-e712-4243-bc3b-2658696a1bae service nova] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] No waiting events found dispatching network-vif-plugged-58c44c97-32fa-4f2f-9313-6fee30a10aec {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1028.986367] env[61243]: WARNING nova.compute.manager [req-3157c1cb-7ce9-4ae2-ab7e-ac4c2aa6aad1 req-ac3632d3-e712-4243-bc3b-2658696a1bae service nova] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Received unexpected event network-vif-plugged-58c44c97-32fa-4f2f-9313-6fee30a10aec for instance with vm_state building and task_state spawning. [ 1028.986538] env[61243]: DEBUG nova.compute.manager [req-3157c1cb-7ce9-4ae2-ab7e-ac4c2aa6aad1 req-ac3632d3-e712-4243-bc3b-2658696a1bae service nova] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Received event network-changed-58c44c97-32fa-4f2f-9313-6fee30a10aec {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.986696] env[61243]: DEBUG nova.compute.manager [req-3157c1cb-7ce9-4ae2-ab7e-ac4c2aa6aad1 req-ac3632d3-e712-4243-bc3b-2658696a1bae service nova] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Refreshing instance network info cache due to event network-changed-58c44c97-32fa-4f2f-9313-6fee30a10aec. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1028.986888] env[61243]: DEBUG oslo_concurrency.lockutils [req-3157c1cb-7ce9-4ae2-ab7e-ac4c2aa6aad1 req-ac3632d3-e712-4243-bc3b-2658696a1bae service nova] Acquiring lock "refresh_cache-05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.987043] env[61243]: DEBUG oslo_concurrency.lockutils [req-3157c1cb-7ce9-4ae2-ab7e-ac4c2aa6aad1 req-ac3632d3-e712-4243-bc3b-2658696a1bae service nova] Acquired lock "refresh_cache-05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.987211] env[61243]: DEBUG nova.network.neutron [req-3157c1cb-7ce9-4ae2-ab7e-ac4c2aa6aad1 req-ac3632d3-e712-4243-bc3b-2658696a1bae service nova] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Refreshing network info cache for port 58c44c97-32fa-4f2f-9313-6fee30a10aec {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1029.218385] env[61243]: INFO nova.scheduler.client.report [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleted allocation for migration 50690685-915d-4db4-8806-bb464dd29d17 [ 1029.262749] env[61243]: DEBUG oslo_concurrency.lockutils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "refresh_cache-05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.309527] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.309841] env[61243]: DEBUG nova.compute.manager [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Instance network_info: |[{"id": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "address": "fa:16:3e:43:9a:6f", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde6bbdd5-c1", "ovs_interfaceid": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1029.310450] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:9a:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '93c5b7ce-4c84-40bc-884c-b2453e0eee69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'de6bbdd5-c1ee-40d7-a91c-1938f53863a7', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1029.318159] env[61243]: DEBUG oslo.service.loopingcall [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.320481] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1029.320896] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d65d47b6-b336-463a-b6d5-978146d5d2eb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.342293] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1029.342293] env[61243]: value = "task-1339367" [ 1029.342293] env[61243]: _type = "Task" [ 1029.342293] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.346306] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701ed9c8-de00-4f42-a729-570b6d7bb517 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.354797] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339367, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.355744] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e49b1f9d-697b-4c84-9b7d-d69110e22d25 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.385685] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f94673fb-2acb-4c56-ac1b-131ab71e3b6d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.393119] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a271db8-acb8-4fe0-bdda-fa53d536dcbc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.408213] env[61243]: DEBUG nova.compute.provider_tree [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.522400] env[61243]: DEBUG nova.network.neutron [req-3157c1cb-7ce9-4ae2-ab7e-ac4c2aa6aad1 req-ac3632d3-e712-4243-bc3b-2658696a1bae service nova] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1029.609687] env[61243]: DEBUG nova.network.neutron [req-3157c1cb-7ce9-4ae2-ab7e-ac4c2aa6aad1 req-ac3632d3-e712-4243-bc3b-2658696a1bae service nova] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.727366] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bb02a2fa-9a5e-4348-a2d9-12882def03e8 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "5660de64-7152-4832-960e-2deb1c0bc37e" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 11.769s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.756915] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "55f24903-b0b4-4124-afde-f31b2680a53a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.757174] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.852840] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339367, 'name': CreateVM_Task, 'duration_secs': 0.338129} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.853034] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1029.853746] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.853926] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.854308] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1029.854569] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6281b6bf-1d9e-49a5-8892-fd16440a88a9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.859436] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1029.859436] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527cfe37-0da2-f989-8cae-255089214553" [ 1029.859436] env[61243]: _type = "Task" [ 1029.859436] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.866757] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527cfe37-0da2-f989-8cae-255089214553, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.911807] env[61243]: DEBUG nova.scheduler.client.report [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1030.112513] env[61243]: DEBUG oslo_concurrency.lockutils [req-3157c1cb-7ce9-4ae2-ab7e-ac4c2aa6aad1 req-ac3632d3-e712-4243-bc3b-2658696a1bae service nova] Releasing lock "refresh_cache-05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.112905] env[61243]: DEBUG oslo_concurrency.lockutils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "refresh_cache-05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.113092] env[61243]: DEBUG nova.network.neutron [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1030.260160] env[61243]: DEBUG nova.compute.utils [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1030.369988] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527cfe37-0da2-f989-8cae-255089214553, 'name': SearchDatastore_Task, 'duration_secs': 0.008966} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.370338] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.370589] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1030.370837] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.370991] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.371238] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1030.371528] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c215eef-8ef2-41a2-b756-276baf55f202 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.385543] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1030.385927] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1030.386785] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95c5160a-4a68-4244-bb77-5eab676129ac {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.392467] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1030.392467] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52b2347b-137a-759d-dc73-250d3215528e" [ 1030.392467] env[61243]: _type = "Task" [ 1030.392467] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.401500] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52b2347b-137a-759d-dc73-250d3215528e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.416976] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.770s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.420315] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.128s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.421266] env[61243]: INFO nova.compute.claims [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1030.455333] env[61243]: INFO nova.scheduler.client.report [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Deleted allocations for instance 0e03eee3-e7bc-483f-948f-a5b408375e24 [ 1030.649646] env[61243]: DEBUG nova.network.neutron [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1030.764908] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.800650] env[61243]: DEBUG nova.network.neutron [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Updating instance_info_cache with network_info: [{"id": "58c44c97-32fa-4f2f-9313-6fee30a10aec", "address": "fa:16:3e:94:37:d9", "network": {"id": "bce1aecb-e043-4f68-99e2-c6b03c4bb866", "bridge": "br-int", "label": "tempest-ServersTestJSON-652904947-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e641e72e4d2048218ef403456dbb2b82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58c44c97-32", "ovs_interfaceid": "58c44c97-32fa-4f2f-9313-6fee30a10aec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.903269] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52b2347b-137a-759d-dc73-250d3215528e, 'name': SearchDatastore_Task, 'duration_secs': 0.025231} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.904073] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e4f3866-1b89-4aa2-a4cc-f3ff74b3ba8d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.908855] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1030.908855] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523690b3-51de-f475-f54b-4f58c1b19625" [ 1030.908855] env[61243]: _type = "Task" [ 1030.908855] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.916078] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523690b3-51de-f475-f54b-4f58c1b19625, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.967175] env[61243]: DEBUG oslo_concurrency.lockutils [None req-1d37b26f-20cb-4995-abc4-b13ce02b5107 tempest-ServerMetadataNegativeTestJSON-788579819 tempest-ServerMetadataNegativeTestJSON-788579819-project-member] Lock "0e03eee3-e7bc-483f-948f-a5b408375e24" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.843s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.294799] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "2c4217cd-af4a-4d64-a422-73955326df03" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.295143] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "2c4217cd-af4a-4d64-a422-73955326df03" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.295361] env[61243]: DEBUG nova.compute.manager [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Going to confirm migration 3 {{(pid=61243) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1031.302459] env[61243]: DEBUG oslo_concurrency.lockutils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "refresh_cache-05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.302739] env[61243]: DEBUG nova.compute.manager [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Instance network_info: |[{"id": "58c44c97-32fa-4f2f-9313-6fee30a10aec", "address": "fa:16:3e:94:37:d9", "network": {"id": "bce1aecb-e043-4f68-99e2-c6b03c4bb866", "bridge": "br-int", "label": "tempest-ServersTestJSON-652904947-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e641e72e4d2048218ef403456dbb2b82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58c44c97-32", "ovs_interfaceid": "58c44c97-32fa-4f2f-9313-6fee30a10aec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1031.303147] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:37:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '90878b7b-ddb7-4f47-892b-d6e06f73475f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '58c44c97-32fa-4f2f-9313-6fee30a10aec', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1031.310542] env[61243]: DEBUG oslo.service.loopingcall [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1031.310762] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1031.310992] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4c8f8f9-f107-43e8-a7f9-f2c97a708b89 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.333982] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1031.333982] env[61243]: value = "task-1339368" [ 1031.333982] env[61243]: _type = "Task" [ 1031.333982] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.343051] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339368, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.418786] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523690b3-51de-f475-f54b-4f58c1b19625, 'name': SearchDatastore_Task, 'duration_secs': 0.009544} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.419079] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.419360] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 466f6a0a-9bc2-4709-8625-e6fc65b68f80/466f6a0a-9bc2-4709-8625-e6fc65b68f80.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1031.419623] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36e356ea-fb7a-4038-be97-60758489c295 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.425782] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1031.425782] env[61243]: value = "task-1339369" [ 1031.425782] env[61243]: _type = "Task" [ 1031.425782] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.436608] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339369, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.587018] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9157963-7166-4f0e-9632-412204e23033 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.597296] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f349ef2-3852-44cc-b92e-27b46d2967ed {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.630712] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e86c2b2-fcf3-4fdb-92ef-84c5ac0dc503 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.642464] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c285f810-15ee-457f-8d68-0b77f87c4388 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.658270] env[61243]: DEBUG nova.compute.provider_tree [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1031.830116] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "55f24903-b0b4-4124-afde-f31b2680a53a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.830634] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.831102] env[61243]: INFO nova.compute.manager [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Attaching volume 21807d69-846e-400f-b7f6-1d044e77e6c7 to /dev/sdb [ 1031.857030] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339368, 'name': CreateVM_Task, 'duration_secs': 0.304547} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.857030] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1031.857822] env[61243]: DEBUG oslo_concurrency.lockutils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.858182] env[61243]: DEBUG oslo_concurrency.lockutils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.859972] env[61243]: DEBUG oslo_concurrency.lockutils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1031.860453] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58ec6046-428d-4f9c-b054-d492db51ec71 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.868606] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1031.868606] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52faf8b0-8bb8-5180-56fc-12aaf77f4b76" [ 1031.868606] env[61243]: _type = "Task" [ 1031.868606] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.876494] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-691e1fd3-4964-406f-934b-2fc01bfedff0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.888025] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52faf8b0-8bb8-5180-56fc-12aaf77f4b76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.888576] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1963d02-ea11-4030-9fd8-780c0dc9bf8e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.904334] env[61243]: DEBUG nova.virt.block_device [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Updating existing volume attachment record: 13dd73b8-019a-4ba2-8254-9f7c9270cba5 {{(pid=61243) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1031.909572] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "refresh_cache-2c4217cd-af4a-4d64-a422-73955326df03" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.909753] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquired lock "refresh_cache-2c4217cd-af4a-4d64-a422-73955326df03" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.909808] env[61243]: DEBUG nova.network.neutron [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1031.909982] env[61243]: DEBUG nova.objects.instance [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lazy-loading 'info_cache' on Instance uuid 2c4217cd-af4a-4d64-a422-73955326df03 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.938864] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339369, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.138404] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "5660de64-7152-4832-960e-2deb1c0bc37e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.138695] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "5660de64-7152-4832-960e-2deb1c0bc37e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.139246] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "5660de64-7152-4832-960e-2deb1c0bc37e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.139547] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "5660de64-7152-4832-960e-2deb1c0bc37e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.139744] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "5660de64-7152-4832-960e-2deb1c0bc37e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.142551] env[61243]: INFO nova.compute.manager [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Terminating instance [ 1032.145959] env[61243]: DEBUG nova.compute.manager [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1032.146225] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1032.148101] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d01c7e-bd09-4787-9ab8-6713877d34db {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.157642] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1032.157642] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-088f6310-201a-4ef8-8596-dab347f608d5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.161355] env[61243]: DEBUG nova.scheduler.client.report [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.165976] env[61243]: DEBUG oslo_vmware.api [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1032.165976] env[61243]: value = "task-1339371" [ 1032.165976] env[61243]: _type = "Task" [ 1032.165976] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.178460] env[61243]: DEBUG oslo_vmware.api [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339371, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.380601] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52faf8b0-8bb8-5180-56fc-12aaf77f4b76, 'name': SearchDatastore_Task, 'duration_secs': 0.055891} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.380933] env[61243]: DEBUG oslo_concurrency.lockutils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.381524] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1032.381627] env[61243]: DEBUG oslo_concurrency.lockutils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.381779] env[61243]: DEBUG oslo_concurrency.lockutils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.382019] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1032.382262] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-86c9d075-a7d7-4f64-8efd-1bc36054a63a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.391175] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1032.391175] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1032.391777] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c00a0df-adce-4685-82c8-12699be13399 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.397602] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1032.397602] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5261d2e3-850e-9d1f-5cd6-f64d0ac74d77" [ 1032.397602] env[61243]: _type = "Task" [ 1032.397602] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.406060] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5261d2e3-850e-9d1f-5cd6-f64d0ac74d77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.440348] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339369, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.549677} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.440348] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 466f6a0a-9bc2-4709-8625-e6fc65b68f80/466f6a0a-9bc2-4709-8625-e6fc65b68f80.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1032.440348] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1032.440845] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0e7a8dbb-7500-416f-8f71-19c83c435b0c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.451294] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1032.451294] env[61243]: value = "task-1339372" [ 1032.451294] env[61243]: _type = "Task" [ 1032.451294] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.460977] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339372, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.632816] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.632988] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.667898] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.248s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.668462] env[61243]: DEBUG nova.compute.manager [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1032.681857] env[61243]: DEBUG oslo_vmware.api [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339371, 'name': PowerOffVM_Task, 'duration_secs': 0.204321} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.681857] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1032.682628] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1032.682628] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-32317cfe-1d50-40ad-9309-24f8e899d79d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.745448] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1032.745753] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1032.746273] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleting the datastore file [datastore2] 5660de64-7152-4832-960e-2deb1c0bc37e {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1032.746273] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4357a79f-8d9c-4857-9c7c-4a80e99f1e07 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.752272] env[61243]: DEBUG oslo_vmware.api [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1032.752272] env[61243]: value = "task-1339374" [ 1032.752272] env[61243]: _type = "Task" [ 1032.752272] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.760372] env[61243]: DEBUG oslo_vmware.api [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339374, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.910544] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5261d2e3-850e-9d1f-5cd6-f64d0ac74d77, 'name': SearchDatastore_Task, 'duration_secs': 0.03618} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.910884] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd6558c7-1d43-45a7-916c-a45e96d2a2ef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.918191] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1032.918191] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ff2361-7717-3186-02c7-57b705b4df6c" [ 1032.918191] env[61243]: _type = "Task" [ 1032.918191] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.928417] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ff2361-7717-3186-02c7-57b705b4df6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.959958] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339372, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.165245} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.960264] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1032.961079] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e411bb37-b711-4014-a2c3-d704375ed5bd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.983217] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 466f6a0a-9bc2-4709-8625-e6fc65b68f80/466f6a0a-9bc2-4709-8625-e6fc65b68f80.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.983515] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef322a76-6d43-4860-9940-70f5ff2f680a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.005782] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1033.005782] env[61243]: value = "task-1339375" [ 1033.005782] env[61243]: _type = "Task" [ 1033.005782] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.014221] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339375, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.135693] env[61243]: DEBUG nova.compute.manager [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1033.143221] env[61243]: DEBUG nova.network.neutron [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Updating instance_info_cache with network_info: [{"id": "611257cf-50e1-491a-beae-8c942398013d", "address": "fa:16:3e:6a:c8:a1", "network": {"id": "0788f9ac-5a15-4431-a711-9b4f6cf79fbc", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1210957657-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "113713170fb94b8a8d7985b9a02faf77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c7821ea-f92f-4f06-a4cb-05e1186a9d22", "external-id": "nsx-vlan-transportzone-69", "segmentation_id": 69, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap611257cf-50", "ovs_interfaceid": "611257cf-50e1-491a-beae-8c942398013d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.175143] env[61243]: DEBUG nova.compute.utils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1033.176085] env[61243]: DEBUG nova.compute.manager [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1033.177037] env[61243]: DEBUG nova.network.neutron [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1033.225637] env[61243]: DEBUG nova.policy [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '876a16a17dd34418b20be920e53779f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2465299214be4d5f877f340489215d6d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1033.263826] env[61243]: DEBUG oslo_vmware.api [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339374, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.321915} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.264215] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1033.264464] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1033.264707] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1033.264922] env[61243]: INFO nova.compute.manager [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1033.265262] env[61243]: DEBUG oslo.service.loopingcall [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.265511] env[61243]: DEBUG nova.compute.manager [-] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1033.265642] env[61243]: DEBUG nova.network.neutron [-] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1033.428915] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ff2361-7717-3186-02c7-57b705b4df6c, 'name': SearchDatastore_Task, 'duration_secs': 0.021907} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.429218] env[61243]: DEBUG oslo_concurrency.lockutils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.429483] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a/05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1033.429736] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1e532b53-d699-42b9-8a28-2dc180a2eee3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.436672] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1033.436672] env[61243]: value = "task-1339376" [ 1033.436672] env[61243]: _type = "Task" [ 1033.436672] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.445534] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339376, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.515456] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339375, 'name': ReconfigVM_Task, 'duration_secs': 0.324036} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.515746] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 466f6a0a-9bc2-4709-8625-e6fc65b68f80/466f6a0a-9bc2-4709-8625-e6fc65b68f80.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.516400] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eb5e364c-07dd-4ae4-a9c9-c7121b6ab964 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.522544] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1033.522544] env[61243]: value = "task-1339377" [ 1033.522544] env[61243]: _type = "Task" [ 1033.522544] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.532275] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339377, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.646967] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Releasing lock "refresh_cache-2c4217cd-af4a-4d64-a422-73955326df03" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.646967] env[61243]: DEBUG nova.objects.instance [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lazy-loading 'migration_context' on Instance uuid 2c4217cd-af4a-4d64-a422-73955326df03 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.662800] env[61243]: DEBUG nova.compute.manager [req-37bb3ba1-f861-413a-a2a2-f355b64b49c4 req-9fcbc32c-3086-435d-8cd5-235d4bf91788 service nova] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Received event network-vif-deleted-2b065192-815c-492f-8f79-bfa9553ff080 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1033.663161] env[61243]: INFO nova.compute.manager [req-37bb3ba1-f861-413a-a2a2-f355b64b49c4 req-9fcbc32c-3086-435d-8cd5-235d4bf91788 service nova] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Neutron deleted interface 2b065192-815c-492f-8f79-bfa9553ff080; detaching it from the instance and deleting it from the info cache [ 1033.663467] env[61243]: DEBUG nova.network.neutron [req-37bb3ba1-f861-413a-a2a2-f355b64b49c4 req-9fcbc32c-3086-435d-8cd5-235d4bf91788 service nova] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.666219] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.666546] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.667910] env[61243]: INFO nova.compute.claims [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1033.679173] env[61243]: DEBUG nova.compute.manager [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1033.685165] env[61243]: DEBUG nova.network.neutron [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Successfully created port: 00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1033.946482] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339376, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.033310] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339377, 'name': Rename_Task, 'duration_secs': 0.155823} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.033560] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1034.033835] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5aa19705-63a8-4a60-88d7-be07381565f6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.039746] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1034.039746] env[61243]: value = "task-1339378" [ 1034.039746] env[61243]: _type = "Task" [ 1034.039746] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.046681] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339378, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.132335] env[61243]: DEBUG nova.network.neutron [-] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.149377] env[61243]: DEBUG nova.objects.base [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Object Instance<2c4217cd-af4a-4d64-a422-73955326df03> lazy-loaded attributes: info_cache,migration_context {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1034.150458] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5daafa-776b-43ff-9da0-b5d50fbd3f0b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.170194] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba4d66b1-f5bc-4be2-abe7-15bcf71a62e1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.175224] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0baf8be3-29d6-4079-aaf7-9eeeb6c673c5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.181027] env[61243]: DEBUG oslo_vmware.api [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 1034.181027] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528fa773-656e-984a-f1ff-4adee09c597b" [ 1034.181027] env[61243]: _type = "Task" [ 1034.181027] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.193182] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce9596ab-6406-4ad7-8c66-24f536b8a553 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.207994] env[61243]: DEBUG oslo_vmware.api [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528fa773-656e-984a-f1ff-4adee09c597b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.225744] env[61243]: DEBUG nova.compute.manager [req-37bb3ba1-f861-413a-a2a2-f355b64b49c4 req-9fcbc32c-3086-435d-8cd5-235d4bf91788 service nova] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Detach interface failed, port_id=2b065192-815c-492f-8f79-bfa9553ff080, reason: Instance 5660de64-7152-4832-960e-2deb1c0bc37e could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1034.449766] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339376, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.943242} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.450065] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a/05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1034.450297] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1034.450559] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e6065e61-3d84-4859-957c-d17a024428e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.456887] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1034.456887] env[61243]: value = "task-1339380" [ 1034.456887] env[61243]: _type = "Task" [ 1034.456887] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.464749] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339380, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.550465] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339378, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.635229] env[61243]: INFO nova.compute.manager [-] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Took 1.37 seconds to deallocate network for instance. [ 1034.692701] env[61243]: DEBUG nova.compute.manager [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1034.695095] env[61243]: DEBUG oslo_vmware.api [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528fa773-656e-984a-f1ff-4adee09c597b, 'name': SearchDatastore_Task, 'duration_secs': 0.028247} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.695673] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.719099] env[61243]: DEBUG nova.virt.hardware [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1034.719377] env[61243]: DEBUG nova.virt.hardware [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1034.719544] env[61243]: DEBUG nova.virt.hardware [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1034.719734] env[61243]: DEBUG nova.virt.hardware [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1034.719908] env[61243]: DEBUG nova.virt.hardware [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1034.720080] env[61243]: DEBUG nova.virt.hardware [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1034.720309] env[61243]: DEBUG nova.virt.hardware [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1034.720475] env[61243]: DEBUG nova.virt.hardware [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1034.720649] env[61243]: DEBUG nova.virt.hardware [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1034.720814] env[61243]: DEBUG nova.virt.hardware [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1034.720994] env[61243]: DEBUG nova.virt.hardware [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1034.721916] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f99307f-8098-4803-9896-f1c7f6340d71 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.730876] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca57745d-d878-4e5a-a102-45b251222f2d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.831248] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b9b0a2-cd96-45a0-ab53-5e014b5baba7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.838729] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb53f99-6952-4607-91a6-82dd30ccfba8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.867504] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4655905-0d29-49de-b795-52858b580efe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.874396] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db61a86-6ed9-493b-a8ae-1d43826a82b6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.888800] env[61243]: DEBUG nova.compute.provider_tree [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1034.967249] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339380, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064164} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.967528] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1034.968280] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc320f7-c629-4c72-ab25-26bcdbd094c4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.989713] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a/05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1034.990498] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54fb15f2-669d-4200-a41d-715ee895a2b8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.008803] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1035.008803] env[61243]: value = "task-1339381" [ 1035.008803] env[61243]: _type = "Task" [ 1035.008803] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.016485] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339381, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.049121] env[61243]: DEBUG oslo_vmware.api [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339378, 'name': PowerOnVM_Task, 'duration_secs': 0.66581} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.049477] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1035.049613] env[61243]: INFO nova.compute.manager [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Took 8.91 seconds to spawn the instance on the hypervisor. [ 1035.049795] env[61243]: DEBUG nova.compute.manager [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1035.050601] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262fae91-04eb-4266-9a7a-a12e2bb067ef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.142086] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.146183] env[61243]: DEBUG nova.network.neutron [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Successfully updated port: 00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1035.410022] env[61243]: ERROR nova.scheduler.client.report [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [req-dcabc8b5-9a2a-410c-ae54-0794955a74e0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-dcabc8b5-9a2a-410c-ae54-0794955a74e0"}]} [ 1035.425562] env[61243]: DEBUG nova.scheduler.client.report [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1035.440760] env[61243]: DEBUG nova.scheduler.client.report [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1035.441063] env[61243]: DEBUG nova.compute.provider_tree [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1035.452627] env[61243]: DEBUG nova.scheduler.client.report [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1035.471941] env[61243]: DEBUG nova.scheduler.client.report [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1035.522072] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339381, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.570868] env[61243]: INFO nova.compute.manager [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Took 14.55 seconds to build instance. [ 1035.629813] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa509d5-6757-4bd9-b1ed-f9c983f6752b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.637372] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b64a9ea7-9c9d-4bf1-8484-45919fc3baea {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.668731] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.668887] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.669060] env[61243]: DEBUG nova.network.neutron [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1035.671586] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ae41fac-55bd-4d8d-8d90-f4e3914f21c9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.679805] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e9c1e8-de16-4a28-be86-3b514833bb3d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.694190] env[61243]: DEBUG nova.compute.provider_tree [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1035.697437] env[61243]: DEBUG nova.compute.manager [req-fbf7fa55-1369-41c5-965a-ccbfdbea9827 req-bf9e82de-b5a3-4275-a342-f6e7271e59b4 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Received event network-vif-plugged-00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1035.697678] env[61243]: DEBUG oslo_concurrency.lockutils [req-fbf7fa55-1369-41c5-965a-ccbfdbea9827 req-bf9e82de-b5a3-4275-a342-f6e7271e59b4 service nova] Acquiring lock "486ebf65-4376-424c-9deb-f1ef70e44be3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.697894] env[61243]: DEBUG oslo_concurrency.lockutils [req-fbf7fa55-1369-41c5-965a-ccbfdbea9827 req-bf9e82de-b5a3-4275-a342-f6e7271e59b4 service nova] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.698119] env[61243]: DEBUG oslo_concurrency.lockutils [req-fbf7fa55-1369-41c5-965a-ccbfdbea9827 req-bf9e82de-b5a3-4275-a342-f6e7271e59b4 service nova] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.698255] env[61243]: DEBUG nova.compute.manager [req-fbf7fa55-1369-41c5-965a-ccbfdbea9827 req-bf9e82de-b5a3-4275-a342-f6e7271e59b4 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] No waiting events found dispatching network-vif-plugged-00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1035.698470] env[61243]: WARNING nova.compute.manager [req-fbf7fa55-1369-41c5-965a-ccbfdbea9827 req-bf9e82de-b5a3-4275-a342-f6e7271e59b4 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Received unexpected event network-vif-plugged-00f9d661-7722-4165-8933-c025ac1f938a for instance with vm_state building and task_state spawning. [ 1035.698658] env[61243]: DEBUG nova.compute.manager [req-fbf7fa55-1369-41c5-965a-ccbfdbea9827 req-bf9e82de-b5a3-4275-a342-f6e7271e59b4 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Received event network-changed-00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1035.698915] env[61243]: DEBUG nova.compute.manager [req-fbf7fa55-1369-41c5-965a-ccbfdbea9827 req-bf9e82de-b5a3-4275-a342-f6e7271e59b4 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Refreshing instance network info cache due to event network-changed-00f9d661-7722-4165-8933-c025ac1f938a. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1035.699085] env[61243]: DEBUG oslo_concurrency.lockutils [req-fbf7fa55-1369-41c5-965a-ccbfdbea9827 req-bf9e82de-b5a3-4275-a342-f6e7271e59b4 service nova] Acquiring lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.022292] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339381, 'name': ReconfigVM_Task, 'duration_secs': 0.94495} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.022631] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a/05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1036.023290] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ef8e884e-e616-4dbf-8057-3e4235584f4f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.029322] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1036.029322] env[61243]: value = "task-1339382" [ 1036.029322] env[61243]: _type = "Task" [ 1036.029322] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.037313] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339382, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.073047] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f2703a7c-adb0-4193-b2c3-6df8569a937e tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.062s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.203271] env[61243]: DEBUG nova.network.neutron [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1036.235109] env[61243]: DEBUG nova.scheduler.client.report [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 143 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1036.235427] env[61243]: DEBUG nova.compute.provider_tree [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 143 to 144 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1036.235607] env[61243]: DEBUG nova.compute.provider_tree [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1036.339814] env[61243]: DEBUG nova.network.neutron [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Updating instance_info_cache with network_info: [{"id": "00f9d661-7722-4165-8933-c025ac1f938a", "address": "fa:16:3e:e9:b9:d7", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00f9d661-77", "ovs_interfaceid": "00f9d661-7722-4165-8933-c025ac1f938a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.462566] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Volume attach. Driver type: vmdk {{(pid=61243) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1036.462833] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285776', 'volume_id': '21807d69-846e-400f-b7f6-1d044e77e6c7', 'name': 'volume-21807d69-846e-400f-b7f6-1d044e77e6c7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '55f24903-b0b4-4124-afde-f31b2680a53a', 'attached_at': '', 'detached_at': '', 'volume_id': '21807d69-846e-400f-b7f6-1d044e77e6c7', 'serial': '21807d69-846e-400f-b7f6-1d044e77e6c7'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1036.463800] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01002b92-e674-40e8-a27c-f42a6bbd0d69 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.480459] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e480d5f-5bdc-4836-a4ca-3a03089992f1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.511868] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] volume-21807d69-846e-400f-b7f6-1d044e77e6c7/volume-21807d69-846e-400f-b7f6-1d044e77e6c7.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1036.511868] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b86af6f-5b40-4c01-a20f-698205603308 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.527746] env[61243]: DEBUG nova.compute.manager [req-d4100957-ac12-43fe-a14f-856825bbdc1f req-d9653237-d9bb-4f55-afca-f157cf0dff30 service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Received event network-changed-de6bbdd5-c1ee-40d7-a91c-1938f53863a7 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1036.527961] env[61243]: DEBUG nova.compute.manager [req-d4100957-ac12-43fe-a14f-856825bbdc1f req-d9653237-d9bb-4f55-afca-f157cf0dff30 service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Refreshing instance network info cache due to event network-changed-de6bbdd5-c1ee-40d7-a91c-1938f53863a7. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1036.528327] env[61243]: DEBUG oslo_concurrency.lockutils [req-d4100957-ac12-43fe-a14f-856825bbdc1f req-d9653237-d9bb-4f55-afca-f157cf0dff30 service nova] Acquiring lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.528373] env[61243]: DEBUG oslo_concurrency.lockutils [req-d4100957-ac12-43fe-a14f-856825bbdc1f req-d9653237-d9bb-4f55-afca-f157cf0dff30 service nova] Acquired lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.528512] env[61243]: DEBUG nova.network.neutron [req-d4100957-ac12-43fe-a14f-856825bbdc1f req-d9653237-d9bb-4f55-afca-f157cf0dff30 service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Refreshing network info cache for port de6bbdd5-c1ee-40d7-a91c-1938f53863a7 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1036.531209] env[61243]: DEBUG oslo_vmware.api [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 1036.531209] env[61243]: value = "task-1339383" [ 1036.531209] env[61243]: _type = "Task" [ 1036.531209] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.545824] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339382, 'name': Rename_Task, 'duration_secs': 0.261478} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.550189] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1036.551222] env[61243]: DEBUG oslo_vmware.api [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339383, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.551493] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d7142e9e-a7a1-411a-95c7-69f197ed13e8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.557999] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1036.557999] env[61243]: value = "task-1339384" [ 1036.557999] env[61243]: _type = "Task" [ 1036.557999] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.566281] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339384, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.740965] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.074s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.741519] env[61243]: DEBUG nova.compute.manager [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1036.744143] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.048s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.842633] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Releasing lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.842902] env[61243]: DEBUG nova.compute.manager [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Instance network_info: |[{"id": "00f9d661-7722-4165-8933-c025ac1f938a", "address": "fa:16:3e:e9:b9:d7", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00f9d661-77", "ovs_interfaceid": "00f9d661-7722-4165-8933-c025ac1f938a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1036.843286] env[61243]: DEBUG oslo_concurrency.lockutils [req-fbf7fa55-1369-41c5-965a-ccbfdbea9827 req-bf9e82de-b5a3-4275-a342-f6e7271e59b4 service nova] Acquired lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.843487] env[61243]: DEBUG nova.network.neutron [req-fbf7fa55-1369-41c5-965a-ccbfdbea9827 req-bf9e82de-b5a3-4275-a342-f6e7271e59b4 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Refreshing network info cache for port 00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1036.844800] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e9:b9:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d413776-9a8c-4afd-856f-10dbb062ca95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '00f9d661-7722-4165-8933-c025ac1f938a', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1036.852380] env[61243]: DEBUG oslo.service.loopingcall [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.855358] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1036.855869] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-812b55bc-370a-4e43-b33e-5b41691850e8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.875666] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1036.875666] env[61243]: value = "task-1339385" [ 1036.875666] env[61243]: _type = "Task" [ 1036.875666] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.883382] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339385, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.049673] env[61243]: DEBUG oslo_vmware.api [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339383, 'name': ReconfigVM_Task, 'duration_secs': 0.511426} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.050094] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Reconfigured VM instance instance-0000005b to attach disk [datastore2] volume-21807d69-846e-400f-b7f6-1d044e77e6c7/volume-21807d69-846e-400f-b7f6-1d044e77e6c7.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1037.057152] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87620fbe-d98e-41d5-8a01-5cbc836ace7b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.083013] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339384, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.083438] env[61243]: DEBUG oslo_vmware.api [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 1037.083438] env[61243]: value = "task-1339386" [ 1037.083438] env[61243]: _type = "Task" [ 1037.083438] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.093943] env[61243]: DEBUG oslo_vmware.api [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339386, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.102617] env[61243]: DEBUG nova.network.neutron [req-fbf7fa55-1369-41c5-965a-ccbfdbea9827 req-bf9e82de-b5a3-4275-a342-f6e7271e59b4 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Updated VIF entry in instance network info cache for port 00f9d661-7722-4165-8933-c025ac1f938a. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1037.103053] env[61243]: DEBUG nova.network.neutron [req-fbf7fa55-1369-41c5-965a-ccbfdbea9827 req-bf9e82de-b5a3-4275-a342-f6e7271e59b4 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Updating instance_info_cache with network_info: [{"id": "00f9d661-7722-4165-8933-c025ac1f938a", "address": "fa:16:3e:e9:b9:d7", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00f9d661-77", "ovs_interfaceid": "00f9d661-7722-4165-8933-c025ac1f938a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.112382] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.113083] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.247238] env[61243]: DEBUG nova.compute.utils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1037.253320] env[61243]: DEBUG nova.compute.manager [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1037.253320] env[61243]: DEBUG nova.network.neutron [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1037.324031] env[61243]: DEBUG nova.policy [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '785fd146024849efa0b67b1a7f731393', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9db958446334496bab832b95f2363f2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1037.341379] env[61243]: DEBUG nova.network.neutron [req-d4100957-ac12-43fe-a14f-856825bbdc1f req-d9653237-d9bb-4f55-afca-f157cf0dff30 service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Updated VIF entry in instance network info cache for port de6bbdd5-c1ee-40d7-a91c-1938f53863a7. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1037.341746] env[61243]: DEBUG nova.network.neutron [req-d4100957-ac12-43fe-a14f-856825bbdc1f req-d9653237-d9bb-4f55-afca-f157cf0dff30 service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Updating instance_info_cache with network_info: [{"id": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "address": "fa:16:3e:43:9a:6f", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde6bbdd5-c1", "ovs_interfaceid": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.388478] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339385, 'name': CreateVM_Task, 'duration_secs': 0.362859} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.388652] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1037.389387] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.390532] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.390532] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1037.390923] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97219be4-853d-4508-9192-172740f08037 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.397214] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1037.397214] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52945fe9-6461-db8f-b806-de194a3893fb" [ 1037.397214] env[61243]: _type = "Task" [ 1037.397214] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.404714] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52945fe9-6461-db8f-b806-de194a3893fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.439476] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db177cc2-a226-4bf9-b02b-06b5f21fe0c6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.447009] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34506db4-8dfb-4fc6-8e13-2e904679f9d1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.478547] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04f3110-0240-4c29-8c82-1dc3e4ab81a0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.486687] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e99b00d-8ed3-4672-9e4a-7a2383bb59c4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.500059] env[61243]: DEBUG nova.compute.provider_tree [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.578436] env[61243]: DEBUG oslo_vmware.api [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339384, 'name': PowerOnVM_Task, 'duration_secs': 0.593085} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.578728] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1037.578969] env[61243]: INFO nova.compute.manager [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Took 9.17 seconds to spawn the instance on the hypervisor. [ 1037.579273] env[61243]: DEBUG nova.compute.manager [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1037.580195] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-782669a7-64e6-4e24-9dd4-73668ed7736b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.598250] env[61243]: DEBUG oslo_vmware.api [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339386, 'name': ReconfigVM_Task, 'duration_secs': 0.149902} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.598624] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285776', 'volume_id': '21807d69-846e-400f-b7f6-1d044e77e6c7', 'name': 'volume-21807d69-846e-400f-b7f6-1d044e77e6c7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '55f24903-b0b4-4124-afde-f31b2680a53a', 'attached_at': '', 'detached_at': '', 'volume_id': '21807d69-846e-400f-b7f6-1d044e77e6c7', 'serial': '21807d69-846e-400f-b7f6-1d044e77e6c7'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1037.606033] env[61243]: DEBUG oslo_concurrency.lockutils [req-fbf7fa55-1369-41c5-965a-ccbfdbea9827 req-bf9e82de-b5a3-4275-a342-f6e7271e59b4 service nova] Releasing lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.607829] env[61243]: DEBUG nova.network.neutron [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Successfully created port: a968f6e7-b39a-4dd8-abf8-bf0919565a72 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1037.614718] env[61243]: DEBUG nova.compute.manager [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1037.754263] env[61243]: DEBUG nova.compute.manager [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1037.844526] env[61243]: DEBUG oslo_concurrency.lockutils [req-d4100957-ac12-43fe-a14f-856825bbdc1f req-d9653237-d9bb-4f55-afca-f157cf0dff30 service nova] Releasing lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.908128] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52945fe9-6461-db8f-b806-de194a3893fb, 'name': SearchDatastore_Task, 'duration_secs': 0.009113} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.908444] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.908686] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1037.908965] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.909360] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.909414] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1037.909714] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c730af86-45ce-4b56-b313-e0126f50a5da {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.917020] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1037.917208] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1037.917894] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-625bc23a-b981-4723-9a0d-b7f1ebcb6ca2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.922894] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1037.922894] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52dc5301-9af4-56d3-466d-045f38d868c3" [ 1037.922894] env[61243]: _type = "Task" [ 1037.922894] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.930460] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52dc5301-9af4-56d3-466d-045f38d868c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.004439] env[61243]: DEBUG nova.scheduler.client.report [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1038.104311] env[61243]: INFO nova.compute.manager [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Took 17.01 seconds to build instance. [ 1038.139779] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.433094] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52dc5301-9af4-56d3-466d-045f38d868c3, 'name': SearchDatastore_Task, 'duration_secs': 0.008373} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.433939] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c352bd49-bd41-4783-b02d-63d87a055ba5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.439133] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1038.439133] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524032bd-7b4c-1689-5302-6aad998645a3" [ 1038.439133] env[61243]: _type = "Task" [ 1038.439133] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.446523] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524032bd-7b4c-1689-5302-6aad998645a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.606932] env[61243]: DEBUG oslo_concurrency.lockutils [None req-05e30e55-d567-43b4-b731-012ae85d651c tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.531s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.640207] env[61243]: DEBUG nova.objects.instance [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lazy-loading 'flavor' on Instance uuid 55f24903-b0b4-4124-afde-f31b2680a53a {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.764093] env[61243]: DEBUG nova.compute.manager [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1038.788944] env[61243]: DEBUG nova.virt.hardware [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1038.789663] env[61243]: DEBUG nova.virt.hardware [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1038.789856] env[61243]: DEBUG nova.virt.hardware [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1038.790076] env[61243]: DEBUG nova.virt.hardware [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1038.790237] env[61243]: DEBUG nova.virt.hardware [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1038.790391] env[61243]: DEBUG nova.virt.hardware [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1038.790608] env[61243]: DEBUG nova.virt.hardware [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1038.790779] env[61243]: DEBUG nova.virt.hardware [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1038.790957] env[61243]: DEBUG nova.virt.hardware [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1038.791141] env[61243]: DEBUG nova.virt.hardware [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1038.791330] env[61243]: DEBUG nova.virt.hardware [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1038.792655] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b47aec99-c788-42a5-b37e-7113db7e0c0a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.800811] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4fe8874-655d-43df-968d-0996efc432be {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.949731] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]524032bd-7b4c-1689-5302-6aad998645a3, 'name': SearchDatastore_Task, 'duration_secs': 0.020316} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.950034] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.950313] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 486ebf65-4376-424c-9deb-f1ef70e44be3/486ebf65-4376-424c-9deb-f1ef70e44be3.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1038.950584] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84ad9977-f7ec-4f54-92a6-25142b3eef67 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.958430] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1038.958430] env[61243]: value = "task-1339387" [ 1038.958430] env[61243]: _type = "Task" [ 1038.958430] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.967381] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339387, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.014019] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.270s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.017186] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.875s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.020768] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.020768] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.880s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.021224] env[61243]: INFO nova.compute.claims [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1039.045485] env[61243]: INFO nova.scheduler.client.report [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleted allocations for instance 5660de64-7152-4832-960e-2deb1c0bc37e [ 1039.082752] env[61243]: DEBUG nova.compute.manager [req-41182306-c7de-4e40-9014-e43e9f89cd51 req-155d6dfa-60fa-40bc-a5fe-a87ed4988050 service nova] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Received event network-vif-plugged-a968f6e7-b39a-4dd8-abf8-bf0919565a72 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.082987] env[61243]: DEBUG oslo_concurrency.lockutils [req-41182306-c7de-4e40-9014-e43e9f89cd51 req-155d6dfa-60fa-40bc-a5fe-a87ed4988050 service nova] Acquiring lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.083258] env[61243]: DEBUG oslo_concurrency.lockutils [req-41182306-c7de-4e40-9014-e43e9f89cd51 req-155d6dfa-60fa-40bc-a5fe-a87ed4988050 service nova] Lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.083390] env[61243]: DEBUG oslo_concurrency.lockutils [req-41182306-c7de-4e40-9014-e43e9f89cd51 req-155d6dfa-60fa-40bc-a5fe-a87ed4988050 service nova] Lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.084032] env[61243]: DEBUG nova.compute.manager [req-41182306-c7de-4e40-9014-e43e9f89cd51 req-155d6dfa-60fa-40bc-a5fe-a87ed4988050 service nova] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] No waiting events found dispatching network-vif-plugged-a968f6e7-b39a-4dd8-abf8-bf0919565a72 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1039.084032] env[61243]: WARNING nova.compute.manager [req-41182306-c7de-4e40-9014-e43e9f89cd51 req-155d6dfa-60fa-40bc-a5fe-a87ed4988050 service nova] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Received unexpected event network-vif-plugged-a968f6e7-b39a-4dd8-abf8-bf0919565a72 for instance with vm_state building and task_state spawning. [ 1039.147508] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c9517dff-dcd7-4833-a6a5-fcd8339e2d62 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.317s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.320404] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.320574] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.320720] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.320919] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.321127] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.323494] env[61243]: INFO nova.compute.manager [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Terminating instance [ 1039.325582] env[61243]: DEBUG nova.compute.manager [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1039.325781] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1039.326685] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0664f8d6-37af-4767-840a-b1e15bbcdcd1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.336352] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1039.336699] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72a9ced5-c5c3-465d-bfd4-dfa9c40509e8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.344719] env[61243]: DEBUG oslo_vmware.api [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1039.344719] env[61243]: value = "task-1339388" [ 1039.344719] env[61243]: _type = "Task" [ 1039.344719] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.355920] env[61243]: DEBUG oslo_vmware.api [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339388, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.468358] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339387, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488626} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.468644] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 486ebf65-4376-424c-9deb-f1ef70e44be3/486ebf65-4376-424c-9deb-f1ef70e44be3.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1039.468869] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1039.469365] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-190378cb-4e9b-4531-a236-1c9a16a8114d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.475982] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1039.475982] env[61243]: value = "task-1339389" [ 1039.475982] env[61243]: _type = "Task" [ 1039.475982] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.483929] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339389, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.556261] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0d9e43bb-f437-4779-800a-f596132d761a tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "5660de64-7152-4832-960e-2deb1c0bc37e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.417s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.584340] env[61243]: INFO nova.scheduler.client.report [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Deleted allocation for migration 8403be2c-7c0b-4e7f-b48e-ad74b99819e9 [ 1039.596319] env[61243]: DEBUG nova.network.neutron [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Successfully updated port: a968f6e7-b39a-4dd8-abf8-bf0919565a72 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1039.601391] env[61243]: DEBUG oslo_concurrency.lockutils [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "5f1c246d-2732-4da4-a00b-b92e2990aa67" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.601663] env[61243]: DEBUG oslo_concurrency.lockutils [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "5f1c246d-2732-4da4-a00b-b92e2990aa67" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.854766] env[61243]: DEBUG oslo_vmware.api [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339388, 'name': PowerOffVM_Task, 'duration_secs': 0.264295} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.855096] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1039.855271] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1039.855527] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e01ab335-3855-4e27-bcf5-1614353accb6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.986455] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339389, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068025} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.986883] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1039.987551] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4606306a-187b-4ce4-9082-27dd8bf208d2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.015123] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 486ebf65-4376-424c-9deb-f1ef70e44be3/486ebf65-4376-424c-9deb-f1ef70e44be3.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1040.015461] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee3180cd-3c62-4555-958c-a5f1ee7b0355 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.036529] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1040.036529] env[61243]: value = "task-1339391" [ 1040.036529] env[61243]: _type = "Task" [ 1040.036529] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.044394] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339391, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.091217] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "2c4217cd-af4a-4d64-a422-73955326df03" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.796s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.098228] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "refresh_cache-bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.098403] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "refresh_cache-bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.098578] env[61243]: DEBUG nova.network.neutron [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1040.106448] env[61243]: DEBUG nova.compute.utils [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1040.176331] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e5fdc49-e2e6-4210-b0c8-be8c3305cc13 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.184155] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f1dd032-1e2c-466f-bda8-96732b0f7da3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.213188] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd457b0-a68e-40f4-b3f7-e574438445a1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.220367] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddbc6f64-5d66-484e-8df2-1da08e83d1fa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.234836] env[61243]: DEBUG nova.compute.provider_tree [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.547018] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339391, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.609410] env[61243]: DEBUG oslo_concurrency.lockutils [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "5f1c246d-2732-4da4-a00b-b92e2990aa67" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.634303] env[61243]: DEBUG nova.network.neutron [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1040.738349] env[61243]: DEBUG nova.scheduler.client.report [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1040.776908] env[61243]: DEBUG nova.network.neutron [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updating instance_info_cache with network_info: [{"id": "a968f6e7-b39a-4dd8-abf8-bf0919565a72", "address": "fa:16:3e:80:14:77", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa968f6e7-b3", "ovs_interfaceid": "a968f6e7-b39a-4dd8-abf8-bf0919565a72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.851021] env[61243]: DEBUG nova.compute.manager [req-a5e2cd2a-8d48-46e3-b328-7ea39d20afa5 req-91984b0f-dd2b-4aad-85a7-e6de6a9d4a09 service nova] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Received event network-changed-a968f6e7-b39a-4dd8-abf8-bf0919565a72 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1040.851021] env[61243]: DEBUG nova.compute.manager [req-a5e2cd2a-8d48-46e3-b328-7ea39d20afa5 req-91984b0f-dd2b-4aad-85a7-e6de6a9d4a09 service nova] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Refreshing instance network info cache due to event network-changed-a968f6e7-b39a-4dd8-abf8-bf0919565a72. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1040.851021] env[61243]: DEBUG oslo_concurrency.lockutils [req-a5e2cd2a-8d48-46e3-b328-7ea39d20afa5 req-91984b0f-dd2b-4aad-85a7-e6de6a9d4a09 service nova] Acquiring lock "refresh_cache-bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.045245] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "55f24903-b0b4-4124-afde-f31b2680a53a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.045570] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.050260] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339391, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.245267] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.226s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.245792] env[61243]: DEBUG nova.compute.manager [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1041.279833] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "refresh_cache-bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.280235] env[61243]: DEBUG nova.compute.manager [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Instance network_info: |[{"id": "a968f6e7-b39a-4dd8-abf8-bf0919565a72", "address": "fa:16:3e:80:14:77", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa968f6e7-b3", "ovs_interfaceid": "a968f6e7-b39a-4dd8-abf8-bf0919565a72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1041.280591] env[61243]: DEBUG oslo_concurrency.lockutils [req-a5e2cd2a-8d48-46e3-b328-7ea39d20afa5 req-91984b0f-dd2b-4aad-85a7-e6de6a9d4a09 service nova] Acquired lock "refresh_cache-bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.280800] env[61243]: DEBUG nova.network.neutron [req-a5e2cd2a-8d48-46e3-b328-7ea39d20afa5 req-91984b0f-dd2b-4aad-85a7-e6de6a9d4a09 service nova] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Refreshing network info cache for port a968f6e7-b39a-4dd8-abf8-bf0919565a72 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1041.282074] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:14:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a071ecf4-e713-4f97-9271-8c17952f6dee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a968f6e7-b39a-4dd8-abf8-bf0919565a72', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1041.289429] env[61243]: DEBUG oslo.service.loopingcall [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.292462] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1041.293263] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-07ed595c-ca9f-41d1-b35d-6a3ff773383e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.315782] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1041.315782] env[61243]: value = "task-1339392" [ 1041.315782] env[61243]: _type = "Task" [ 1041.315782] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.323325] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "2c4217cd-af4a-4d64-a422-73955326df03" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.323575] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "2c4217cd-af4a-4d64-a422-73955326df03" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.323820] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "2c4217cd-af4a-4d64-a422-73955326df03-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.324028] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "2c4217cd-af4a-4d64-a422-73955326df03-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.324222] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "2c4217cd-af4a-4d64-a422-73955326df03-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.326385] env[61243]: INFO nova.compute.manager [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Terminating instance [ 1041.328054] env[61243]: DEBUG nova.compute.manager [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1041.328279] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1041.328764] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339392, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.329508] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666998a1-6ef1-404d-8c66-84dd514840bb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.337979] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1041.340166] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7f2d80a-7469-4fb8-bda9-8c0d4b839853 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.346314] env[61243]: DEBUG oslo_vmware.api [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 1041.346314] env[61243]: value = "task-1339393" [ 1041.346314] env[61243]: _type = "Task" [ 1041.346314] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.354228] env[61243]: DEBUG oslo_vmware.api [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339393, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.552141] env[61243]: DEBUG nova.compute.utils [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1041.553828] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339391, 'name': ReconfigVM_Task, 'duration_secs': 1.219169} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.554194] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 486ebf65-4376-424c-9deb-f1ef70e44be3/486ebf65-4376-424c-9deb-f1ef70e44be3.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1041.554965] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d9e66b61-948b-44f9-835e-241e54b6f126 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.561925] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1041.561925] env[61243]: value = "task-1339394" [ 1041.561925] env[61243]: _type = "Task" [ 1041.561925] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.572226] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339394, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.679954] env[61243]: DEBUG oslo_concurrency.lockutils [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "5f1c246d-2732-4da4-a00b-b92e2990aa67" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.680233] env[61243]: DEBUG oslo_concurrency.lockutils [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "5f1c246d-2732-4da4-a00b-b92e2990aa67" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.680467] env[61243]: INFO nova.compute.manager [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Attaching volume d494ae0c-78b7-4745-85ac-83f826afe53a to /dev/sdb [ 1041.713409] env[61243]: DEBUG nova.network.neutron [req-a5e2cd2a-8d48-46e3-b328-7ea39d20afa5 req-91984b0f-dd2b-4aad-85a7-e6de6a9d4a09 service nova] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updated VIF entry in instance network info cache for port a968f6e7-b39a-4dd8-abf8-bf0919565a72. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1041.714077] env[61243]: DEBUG nova.network.neutron [req-a5e2cd2a-8d48-46e3-b328-7ea39d20afa5 req-91984b0f-dd2b-4aad-85a7-e6de6a9d4a09 service nova] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updating instance_info_cache with network_info: [{"id": "a968f6e7-b39a-4dd8-abf8-bf0919565a72", "address": "fa:16:3e:80:14:77", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa968f6e7-b3", "ovs_interfaceid": "a968f6e7-b39a-4dd8-abf8-bf0919565a72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.717942] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1f62e8-2677-408d-874b-84c3e2ab63f4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.727576] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7223f24f-9c2a-4b0f-b101-b5415280e9e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.741733] env[61243]: DEBUG nova.virt.block_device [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Updating existing volume attachment record: fbfc3d2c-569a-4b22-86c0-af0f5c364d72 {{(pid=61243) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1041.750290] env[61243]: DEBUG nova.compute.utils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1041.752041] env[61243]: DEBUG nova.compute.manager [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1041.752041] env[61243]: DEBUG nova.network.neutron [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1041.792265] env[61243]: DEBUG nova.policy [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b32f4788825f4a8e82e028cc10347746', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '683c027e4adc4cc4a6b8747a8c8c0de9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1041.827818] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339392, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.856434] env[61243]: DEBUG oslo_vmware.api [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339393, 'name': PowerOffVM_Task, 'duration_secs': 0.201638} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.856628] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1041.856813] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1041.857077] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f1c1db8-8d78-4470-845c-d195134b225b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.927021] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1041.927021] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1041.927021] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Deleting the datastore file [datastore2] 2c4217cd-af4a-4d64-a422-73955326df03 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1041.927021] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d81dcd2f-1f68-4d00-b2e8-e3736745a7f2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.932486] env[61243]: DEBUG oslo_vmware.api [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for the task: (returnval){ [ 1041.932486] env[61243]: value = "task-1339397" [ 1041.932486] env[61243]: _type = "Task" [ 1041.932486] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.940499] env[61243]: DEBUG oslo_vmware.api [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339397, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.058550] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.013s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.071467] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339394, 'name': Rename_Task, 'duration_secs': 0.148441} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.072351] env[61243]: DEBUG nova.network.neutron [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Successfully created port: 5d2d1a82-6eda-413f-9581-9fa9107be4bd {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1042.074304] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1042.074586] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b657b952-0fd8-4ab6-b19d-3dff942641f9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.081145] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1042.081145] env[61243]: value = "task-1339400" [ 1042.081145] env[61243]: _type = "Task" [ 1042.081145] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.095151] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339400, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.175485] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1042.175722] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1042.175904] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleting the datastore file [datastore2] 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1042.176214] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-656238e1-bab8-4066-a9ef-880f66f60ea5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.182441] env[61243]: DEBUG oslo_vmware.api [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1042.182441] env[61243]: value = "task-1339401" [ 1042.182441] env[61243]: _type = "Task" [ 1042.182441] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.190711] env[61243]: DEBUG oslo_vmware.api [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339401, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.223335] env[61243]: DEBUG oslo_concurrency.lockutils [req-a5e2cd2a-8d48-46e3-b328-7ea39d20afa5 req-91984b0f-dd2b-4aad-85a7-e6de6a9d4a09 service nova] Releasing lock "refresh_cache-bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.254543] env[61243]: DEBUG nova.compute.manager [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1042.329659] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339392, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.442527] env[61243]: DEBUG oslo_vmware.api [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Task: {'id': task-1339397, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147196} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.442846] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1042.443133] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1042.443352] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1042.443562] env[61243]: INFO nova.compute.manager [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1042.443875] env[61243]: DEBUG oslo.service.loopingcall [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.444127] env[61243]: DEBUG nova.compute.manager [-] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1042.444197] env[61243]: DEBUG nova.network.neutron [-] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1042.591626] env[61243]: DEBUG oslo_vmware.api [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339400, 'name': PowerOnVM_Task, 'duration_secs': 0.494642} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.591626] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1042.591810] env[61243]: INFO nova.compute.manager [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Took 7.90 seconds to spawn the instance on the hypervisor. [ 1042.591895] env[61243]: DEBUG nova.compute.manager [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1042.593801] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-026178ab-6355-4e84-bf65-06733b563ff8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.693204] env[61243]: DEBUG oslo_vmware.api [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339401, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148133} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.693204] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1042.693204] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1042.693204] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1042.693204] env[61243]: INFO nova.compute.manager [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Took 3.37 seconds to destroy the instance on the hypervisor. [ 1042.693477] env[61243]: DEBUG oslo.service.loopingcall [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.693741] env[61243]: DEBUG nova.compute.manager [-] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1042.693905] env[61243]: DEBUG nova.network.neutron [-] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1042.828239] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339392, 'name': CreateVM_Task, 'duration_secs': 1.083552} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.828467] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1042.829189] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.829416] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.829824] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1042.831182] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72ab8b89-c463-4805-896f-035cc110b77f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.834827] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1042.834827] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5242c282-d9c6-8cb9-bff6-dad804a51ec4" [ 1042.834827] env[61243]: _type = "Task" [ 1042.834827] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.841986] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5242c282-d9c6-8cb9-bff6-dad804a51ec4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.876133] env[61243]: DEBUG nova.compute.manager [req-30495415-3ddd-4bd7-a91e-c9a562433a0e req-fcdfecab-d6b9-4a5e-8eb4-7108faad2830 service nova] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Received event network-vif-deleted-611257cf-50e1-491a-beae-8c942398013d {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.876356] env[61243]: INFO nova.compute.manager [req-30495415-3ddd-4bd7-a91e-c9a562433a0e req-fcdfecab-d6b9-4a5e-8eb4-7108faad2830 service nova] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Neutron deleted interface 611257cf-50e1-491a-beae-8c942398013d; detaching it from the instance and deleting it from the info cache [ 1042.876536] env[61243]: DEBUG nova.network.neutron [req-30495415-3ddd-4bd7-a91e-c9a562433a0e req-fcdfecab-d6b9-4a5e-8eb4-7108faad2830 service nova] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.114510] env[61243]: INFO nova.compute.manager [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Took 18.84 seconds to build instance. [ 1043.126533] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "55f24903-b0b4-4124-afde-f31b2680a53a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.126798] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.127112] env[61243]: INFO nova.compute.manager [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Attaching volume 9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b to /dev/sdc [ 1043.163470] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abdddc04-b656-491d-8ff0-56b0f2cd4df4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.170481] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35a1852-2ecb-4ab3-911d-31b8bf1cbbea {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.183221] env[61243]: DEBUG nova.virt.block_device [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Updating existing volume attachment record: 0e67baf5-4cfd-49b9-b007-92ddab906bdc {{(pid=61243) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1043.202542] env[61243]: DEBUG nova.network.neutron [-] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.266037] env[61243]: DEBUG nova.compute.manager [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1043.290803] env[61243]: DEBUG nova.virt.hardware [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1043.291129] env[61243]: DEBUG nova.virt.hardware [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1043.291329] env[61243]: DEBUG nova.virt.hardware [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1043.291542] env[61243]: DEBUG nova.virt.hardware [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1043.291706] env[61243]: DEBUG nova.virt.hardware [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1043.291850] env[61243]: DEBUG nova.virt.hardware [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1043.292078] env[61243]: DEBUG nova.virt.hardware [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1043.292261] env[61243]: DEBUG nova.virt.hardware [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1043.292450] env[61243]: DEBUG nova.virt.hardware [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1043.292622] env[61243]: DEBUG nova.virt.hardware [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1043.292798] env[61243]: DEBUG nova.virt.hardware [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1043.293659] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df0e5a91-e534-4295-a9bf-ab1767d5dbce {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.300933] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50cea74-0568-4c43-b5f2-61ebd2d1fa34 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.344070] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5242c282-d9c6-8cb9-bff6-dad804a51ec4, 'name': SearchDatastore_Task, 'duration_secs': 0.029938} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.344767] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.345049] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1043.345340] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.345511] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.345705] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1043.345975] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff66734a-d17c-4186-bb96-d6d96de307c1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.355654] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1043.355860] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1043.356792] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44ebce26-c251-4c6e-b89c-73632aeb81d3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.361889] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1043.361889] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5213cb3c-a48a-aada-1073-8e41cd97a5f0" [ 1043.361889] env[61243]: _type = "Task" [ 1043.361889] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.369871] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5213cb3c-a48a-aada-1073-8e41cd97a5f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.379635] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-55d0b33c-cbdc-4fd4-bdb0-b457f5c31022 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.388939] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87081e41-f5f2-4b39-9dd4-40167b170e7e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.417739] env[61243]: DEBUG nova.network.neutron [-] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.419354] env[61243]: DEBUG nova.compute.manager [req-30495415-3ddd-4bd7-a91e-c9a562433a0e req-fcdfecab-d6b9-4a5e-8eb4-7108faad2830 service nova] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Detach interface failed, port_id=611257cf-50e1-491a-beae-8c942398013d, reason: Instance 2c4217cd-af4a-4d64-a422-73955326df03 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1043.600964] env[61243]: DEBUG nova.network.neutron [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Successfully updated port: 5d2d1a82-6eda-413f-9581-9fa9107be4bd {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1043.616503] env[61243]: DEBUG oslo_concurrency.lockutils [None req-cc18faed-b03c-459b-9726-5a958ecb3cc3 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.349s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.705474] env[61243]: INFO nova.compute.manager [-] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Took 1.26 seconds to deallocate network for instance. [ 1043.872073] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5213cb3c-a48a-aada-1073-8e41cd97a5f0, 'name': SearchDatastore_Task, 'duration_secs': 0.008709} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.872861] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01085457-f408-4bb6-8676-399d5ab824b2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.877982] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1043.877982] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d8237b-7be8-5385-8562-bc6e82ca357a" [ 1043.877982] env[61243]: _type = "Task" [ 1043.877982] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.885595] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d8237b-7be8-5385-8562-bc6e82ca357a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.920713] env[61243]: INFO nova.compute.manager [-] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Took 1.23 seconds to deallocate network for instance. [ 1043.935212] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1043.935517] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1043.935663] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Starting heal instance info cache {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1044.103178] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.103449] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.103777] env[61243]: DEBUG nova.network.neutron [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1044.215182] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.215527] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.216234] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.239591] env[61243]: INFO nova.scheduler.client.report [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Deleted allocations for instance 2c4217cd-af4a-4d64-a422-73955326df03 [ 1044.388079] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d8237b-7be8-5385-8562-bc6e82ca357a, 'name': SearchDatastore_Task, 'duration_secs': 0.00862} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.388361] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.388628] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] bb7c810f-91d1-43eb-b0cc-fb4c686b1f29/bb7c810f-91d1-43eb-b0cc-fb4c686b1f29.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1044.388917] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a2798f9-e86f-48d0-8dee-717e93697e91 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.394700] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1044.394700] env[61243]: value = "task-1339404" [ 1044.394700] env[61243]: _type = "Task" [ 1044.394700] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.402213] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339404, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.427440] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.427792] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.428257] env[61243]: DEBUG nova.objects.instance [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lazy-loading 'resources' on Instance uuid 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.638400] env[61243]: DEBUG nova.network.neutron [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1044.752933] env[61243]: DEBUG oslo_concurrency.lockutils [None req-fae915e8-b656-451d-955b-1f3f69ab4e69 tempest-DeleteServersTestJSON-770138440 tempest-DeleteServersTestJSON-770138440-project-member] Lock "2c4217cd-af4a-4d64-a422-73955326df03" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.429s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.791469] env[61243]: DEBUG nova.network.neutron [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Updating instance_info_cache with network_info: [{"id": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "address": "fa:16:3e:a4:7d:cc", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d2d1a82-6e", "ovs_interfaceid": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.904223] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339404, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472981} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.905597] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] bb7c810f-91d1-43eb-b0cc-fb4c686b1f29/bb7c810f-91d1-43eb-b0cc-fb4c686b1f29.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1044.907137] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1044.907611] env[61243]: DEBUG nova.compute.manager [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Received event network-vif-deleted-58c44c97-32fa-4f2f-9313-6fee30a10aec {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1044.907815] env[61243]: DEBUG nova.compute.manager [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Received event network-vif-plugged-5d2d1a82-6eda-413f-9581-9fa9107be4bd {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1044.908010] env[61243]: DEBUG oslo_concurrency.lockutils [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] Acquiring lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.908228] env[61243]: DEBUG oslo_concurrency.lockutils [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] Lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.908400] env[61243]: DEBUG oslo_concurrency.lockutils [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] Lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.908569] env[61243]: DEBUG nova.compute.manager [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] No waiting events found dispatching network-vif-plugged-5d2d1a82-6eda-413f-9581-9fa9107be4bd {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1044.908740] env[61243]: WARNING nova.compute.manager [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Received unexpected event network-vif-plugged-5d2d1a82-6eda-413f-9581-9fa9107be4bd for instance with vm_state building and task_state spawning. [ 1044.908959] env[61243]: DEBUG nova.compute.manager [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Received event network-changed-5d2d1a82-6eda-413f-9581-9fa9107be4bd {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1044.909178] env[61243]: DEBUG nova.compute.manager [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Refreshing instance network info cache due to event network-changed-5d2d1a82-6eda-413f-9581-9fa9107be4bd. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1044.909761] env[61243]: DEBUG oslo_concurrency.lockutils [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] Acquiring lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.909761] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bfe5b1fa-6eff-4a4a-b4cf-ff0af0ea003a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.916083] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1044.916083] env[61243]: value = "task-1339405" [ 1044.916083] env[61243]: _type = "Task" [ 1044.916083] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.924601] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339405, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.055124] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b409ca-5be4-49d3-b16e-d660feeec74b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.065152] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37eab3b9-9a06-4f64-b68f-3ef7252c9aea {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.102199] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f65aec97-b369-4424-90cf-7677a91441c0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.110672] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-debf8e08-d2dc-4b3c-847d-eb4b3b0c1728 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.127816] env[61243]: DEBUG nova.compute.provider_tree [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1045.294742] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.295144] env[61243]: DEBUG nova.compute.manager [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Instance network_info: |[{"id": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "address": "fa:16:3e:a4:7d:cc", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d2d1a82-6e", "ovs_interfaceid": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1045.295549] env[61243]: DEBUG oslo_concurrency.lockutils [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] Acquired lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.295759] env[61243]: DEBUG nova.network.neutron [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Refreshing network info cache for port 5d2d1a82-6eda-413f-9581-9fa9107be4bd {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1045.298170] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a4:7d:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '93c5b7ce-4c84-40bc-884c-b2453e0eee69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5d2d1a82-6eda-413f-9581-9fa9107be4bd', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1045.306092] env[61243]: DEBUG oslo.service.loopingcall [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1045.306769] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1045.307028] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-66209f20-028b-4be5-9ef0-c60041b29204 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.327152] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1045.327152] env[61243]: value = "task-1339406" [ 1045.327152] env[61243]: _type = "Task" [ 1045.327152] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.335478] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339406, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.426456] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339405, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.177007} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.426749] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1045.427568] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc845f14-919d-44ec-ae6e-d622a385cc2a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.450803] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] bb7c810f-91d1-43eb-b0cc-fb4c686b1f29/bb7c810f-91d1-43eb-b0cc-fb4c686b1f29.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1045.453777] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06a55cd0-3e67-4814-905c-ffed4c28c398 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.483547] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1045.483547] env[61243]: value = "task-1339408" [ 1045.483547] env[61243]: _type = "Task" [ 1045.483547] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.491861] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339408, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.650255] env[61243]: ERROR nova.scheduler.client.report [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [req-bde53e5e-31b9-4aa3-babb-45ad6d255b39] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-bde53e5e-31b9-4aa3-babb-45ad6d255b39"}]} [ 1045.665913] env[61243]: DEBUG nova.scheduler.client.report [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1045.679981] env[61243]: DEBUG nova.scheduler.client.report [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1045.680248] env[61243]: DEBUG nova.compute.provider_tree [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1045.690965] env[61243]: DEBUG nova.scheduler.client.report [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1045.708474] env[61243]: DEBUG nova.scheduler.client.report [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1045.838804] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339406, 'name': CreateVM_Task, 'duration_secs': 0.342391} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.841959] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1045.842747] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55ea6e89-c03c-4341-8123-320ad34db8cb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.846094] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.846346] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.846591] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1045.847146] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eeb1ae9a-86ab-4b54-9b1b-bf1d92f9d385 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.854200] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6ef558-2cdd-49ba-8c01-ca05223e32eb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.857185] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1045.857185] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ed2900-3a5e-8313-82f6-87924e5ad729" [ 1045.857185] env[61243]: _type = "Task" [ 1045.857185] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.886403] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff7fae25-cf9d-409d-bc42-28a6ffa0b6f7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.892524] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ed2900-3a5e-8313-82f6-87924e5ad729, 'name': SearchDatastore_Task, 'duration_secs': 0.009203} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.896641] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.896641] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1045.896641] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.896641] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.896641] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1045.898283] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4d605ddf-b827-4c5a-8f21-1a3157b61c6c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.901664] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d126eb2b-88a7-405a-97db-78a169880279 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.916830] env[61243]: DEBUG nova.compute.provider_tree [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1045.925062] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1045.925269] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1045.926825] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7609215c-132e-4793-b9df-ce652ee4f45a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.932416] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1045.932416] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5226b66a-66e9-9731-26ca-5ba4fb640cae" [ 1045.932416] env[61243]: _type = "Task" [ 1045.932416] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.940154] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5226b66a-66e9-9731-26ca-5ba4fb640cae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.993750] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339408, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.041519] env[61243]: DEBUG nova.network.neutron [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Updated VIF entry in instance network info cache for port 5d2d1a82-6eda-413f-9581-9fa9107be4bd. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1046.041904] env[61243]: DEBUG nova.network.neutron [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Updating instance_info_cache with network_info: [{"id": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "address": "fa:16:3e:a4:7d:cc", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d2d1a82-6e", "ovs_interfaceid": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.290695] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Volume attach. Driver type: vmdk {{(pid=61243) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1046.291014] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285780', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'name': 'volume-d494ae0c-78b7-4745-85ac-83f826afe53a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5f1c246d-2732-4da4-a00b-b92e2990aa67', 'attached_at': '', 'detached_at': '', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'serial': 'd494ae0c-78b7-4745-85ac-83f826afe53a'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1046.292191] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2215e3f6-ad87-451b-ae5e-1a5e63fa2c2c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.312136] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d1e7dd-7d78-488e-a1c6-91e12c166cc6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.337774] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] volume-d494ae0c-78b7-4745-85ac-83f826afe53a/volume-d494ae0c-78b7-4745-85ac-83f826afe53a.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.338167] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dc442396-e905-459c-97ef-d665173f8da1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.356388] env[61243]: DEBUG oslo_vmware.api [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1046.356388] env[61243]: value = "task-1339409" [ 1046.356388] env[61243]: _type = "Task" [ 1046.356388] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.364247] env[61243]: DEBUG oslo_vmware.api [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339409, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.442753] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5226b66a-66e9-9731-26ca-5ba4fb640cae, 'name': SearchDatastore_Task, 'duration_secs': 0.01213} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.443440] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03d0c0cc-87bc-40a9-857a-6cf9cbba0662 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.448996] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1046.448996] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52bd88f7-b7be-3182-8e2e-6f7f11f83781" [ 1046.448996] env[61243]: _type = "Task" [ 1046.448996] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.459902] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52bd88f7-b7be-3182-8e2e-6f7f11f83781, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.465046] env[61243]: DEBUG nova.scheduler.client.report [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 146 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1046.465336] env[61243]: DEBUG nova.compute.provider_tree [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 146 to 147 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1046.465905] env[61243]: DEBUG nova.compute.provider_tree [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1046.496621] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339408, 'name': ReconfigVM_Task, 'duration_secs': 0.98942} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.497154] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Reconfigured VM instance instance-00000066 to attach disk [datastore2] bb7c810f-91d1-43eb-b0cc-fb4c686b1f29/bb7c810f-91d1-43eb-b0cc-fb4c686b1f29.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.497824] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93bf12d5-f4e7-4af6-98e5-259cd46ffa3b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.503879] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1046.503879] env[61243]: value = "task-1339410" [ 1046.503879] env[61243]: _type = "Task" [ 1046.503879] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.512242] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339410, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.545599] env[61243]: DEBUG oslo_concurrency.lockutils [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] Releasing lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.545599] env[61243]: DEBUG nova.compute.manager [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Received event network-changed-00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1046.545599] env[61243]: DEBUG nova.compute.manager [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Refreshing instance network info cache due to event network-changed-00f9d661-7722-4165-8933-c025ac1f938a. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1046.545599] env[61243]: DEBUG oslo_concurrency.lockutils [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] Acquiring lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.545775] env[61243]: DEBUG oslo_concurrency.lockutils [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] Acquired lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.546098] env[61243]: DEBUG nova.network.neutron [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Refreshing network info cache for port 00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1046.866162] env[61243]: DEBUG oslo_vmware.api [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339409, 'name': ReconfigVM_Task, 'duration_secs': 0.488591} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.866461] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Reconfigured VM instance instance-0000005e to attach disk [datastore1] volume-d494ae0c-78b7-4745-85ac-83f826afe53a/volume-d494ae0c-78b7-4745-85ac-83f826afe53a.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.871241] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06fc5ffd-2cd1-42b3-be78-a288db1bbeb2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.885762] env[61243]: DEBUG oslo_vmware.api [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1046.885762] env[61243]: value = "task-1339411" [ 1046.885762] env[61243]: _type = "Task" [ 1046.885762] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.893834] env[61243]: DEBUG oslo_vmware.api [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339411, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.960407] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52bd88f7-b7be-3182-8e2e-6f7f11f83781, 'name': SearchDatastore_Task, 'duration_secs': 0.012375} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.960852] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.961213] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 8279496d-4a90-4962-9cc2-5d5e1ee33ff0/8279496d-4a90-4962-9cc2-5d5e1ee33ff0.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1046.961601] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9fcc28af-81f8-4b9e-b4cb-57e5384e5a44 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.969793] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1046.969793] env[61243]: value = "task-1339412" [ 1046.969793] env[61243]: _type = "Task" [ 1046.969793] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.973785] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.546s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.980495] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339412, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.996709] env[61243]: INFO nova.scheduler.client.report [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleted allocations for instance 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a [ 1047.013636] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339410, 'name': Rename_Task, 'duration_secs': 0.21404} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.013995] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1047.014272] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4926a9b-d82a-4b81-930d-c002a921d4cc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.020038] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1047.020038] env[61243]: value = "task-1339413" [ 1047.020038] env[61243]: _type = "Task" [ 1047.020038] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.027725] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339413, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.321065] env[61243]: DEBUG nova.network.neutron [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Updated VIF entry in instance network info cache for port 00f9d661-7722-4165-8933-c025ac1f938a. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1047.321984] env[61243]: DEBUG nova.network.neutron [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Updating instance_info_cache with network_info: [{"id": "00f9d661-7722-4165-8933-c025ac1f938a", "address": "fa:16:3e:e9:b9:d7", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00f9d661-77", "ovs_interfaceid": "00f9d661-7722-4165-8933-c025ac1f938a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.396547] env[61243]: DEBUG oslo_vmware.api [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339411, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.480019] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339412, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475836} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.480434] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 8279496d-4a90-4962-9cc2-5d5e1ee33ff0/8279496d-4a90-4962-9cc2-5d5e1ee33ff0.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1047.480803] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1047.483087] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-09953905-4a52-4063-b137-02831c06bea1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.487968] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1047.487968] env[61243]: value = "task-1339414" [ 1047.487968] env[61243]: _type = "Task" [ 1047.487968] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.502369] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339414, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.508031] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f340578b-6640-4abd-8351-a17469e43143 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.185s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.530829] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339413, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.725804] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Volume attach. Driver type: vmdk {{(pid=61243) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1047.726078] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285781', 'volume_id': '9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b', 'name': 'volume-9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '55f24903-b0b4-4124-afde-f31b2680a53a', 'attached_at': '', 'detached_at': '', 'volume_id': '9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b', 'serial': '9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1047.727710] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30799146-0c53-46de-bcad-62297d20a6ff {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.744106] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3391bed1-2623-40d0-b8b3-e6daaad2f190 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.775613] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] volume-9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b/volume-9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1047.775993] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c0e5139-9ade-4118-84d9-3e2cda665c22 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.797053] env[61243]: DEBUG oslo_vmware.api [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 1047.797053] env[61243]: value = "task-1339415" [ 1047.797053] env[61243]: _type = "Task" [ 1047.797053] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.805411] env[61243]: DEBUG oslo_vmware.api [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339415, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.824694] env[61243]: DEBUG oslo_concurrency.lockutils [req-5ba7a3c2-8e5e-4a5c-bde3-31715d02f481 req-bf7df658-c1c1-4f4c-b99f-0eb0d0b62032 service nova] Releasing lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.896381] env[61243]: DEBUG oslo_vmware.api [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339411, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.001208] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339414, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070728} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.001208] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1048.001208] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d35e2f5-0c2a-460c-9bc5-0fe38ebf8361 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.033350] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 8279496d-4a90-4962-9cc2-5d5e1ee33ff0/8279496d-4a90-4962-9cc2-5d5e1ee33ff0.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1048.037719] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5fbec82-02f9-4dca-abeb-ae1cebc7c196 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.060329] env[61243]: DEBUG oslo_vmware.api [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339413, 'name': PowerOnVM_Task, 'duration_secs': 0.806781} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.060329] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1048.060635] env[61243]: INFO nova.compute.manager [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Took 9.30 seconds to spawn the instance on the hypervisor. [ 1048.060955] env[61243]: DEBUG nova.compute.manager [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1048.061432] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1048.061432] env[61243]: value = "task-1339416" [ 1048.061432] env[61243]: _type = "Task" [ 1048.061432] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.062481] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67721b65-976b-4d93-8916-2f2a37fc2cc9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.076466] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339416, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.309914] env[61243]: DEBUG oslo_vmware.api [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339415, 'name': ReconfigVM_Task, 'duration_secs': 0.384073} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.310499] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Reconfigured VM instance instance-0000005b to attach disk [datastore2] volume-9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b/volume-9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1048.320616] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-09eb8384-f05d-423d-9bac-73be06e3e05f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.345701] env[61243]: DEBUG oslo_vmware.api [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 1048.345701] env[61243]: value = "task-1339417" [ 1048.345701] env[61243]: _type = "Task" [ 1048.345701] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.359220] env[61243]: DEBUG oslo_vmware.api [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339417, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.400066] env[61243]: DEBUG oslo_vmware.api [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339411, 'name': ReconfigVM_Task, 'duration_secs': 1.146492} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.400383] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285780', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'name': 'volume-d494ae0c-78b7-4745-85ac-83f826afe53a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5f1c246d-2732-4da4-a00b-b92e2990aa67', 'attached_at': '', 'detached_at': '', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'serial': 'd494ae0c-78b7-4745-85ac-83f826afe53a'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1048.574324] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339416, 'name': ReconfigVM_Task, 'duration_secs': 0.263985} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.574610] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 8279496d-4a90-4962-9cc2-5d5e1ee33ff0/8279496d-4a90-4962-9cc2-5d5e1ee33ff0.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1048.575252] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1d1cae4f-d4f7-4c89-95b7-483ce86ca273 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.583222] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1048.583222] env[61243]: value = "task-1339418" [ 1048.583222] env[61243]: _type = "Task" [ 1048.583222] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.588921] env[61243]: INFO nova.compute.manager [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Took 14.95 seconds to build instance. [ 1048.596592] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339418, 'name': Rename_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.823413] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "df141b12-1d34-4a47-b6bb-15cb75d266c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.823671] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "df141b12-1d34-4a47-b6bb-15cb75d266c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.856684] env[61243]: DEBUG oslo_vmware.api [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339417, 'name': ReconfigVM_Task, 'duration_secs': 0.135316} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.858904] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285781', 'volume_id': '9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b', 'name': 'volume-9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '55f24903-b0b4-4124-afde-f31b2680a53a', 'attached_at': '', 'detached_at': '', 'volume_id': '9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b', 'serial': '9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1049.091375] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f36d7ac1-1855-4157-89ac-084393c8269f tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.458s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.097645] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339418, 'name': Rename_Task, 'duration_secs': 0.141124} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.097934] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1049.098264] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d8d760ae-11f2-4780-a150-a339eca69c03 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.107394] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1049.107394] env[61243]: value = "task-1339419" [ 1049.107394] env[61243]: _type = "Task" [ 1049.107394] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.115640] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339419, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.333010] env[61243]: DEBUG nova.compute.manager [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1049.442772] env[61243]: DEBUG nova.objects.instance [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lazy-loading 'flavor' on Instance uuid 5f1c246d-2732-4da4-a00b-b92e2990aa67 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.618588] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339419, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.855293] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.856508] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.857318] env[61243]: INFO nova.compute.claims [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1049.909014] env[61243]: DEBUG nova.objects.instance [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lazy-loading 'flavor' on Instance uuid 55f24903-b0b4-4124-afde-f31b2680a53a {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.953943] env[61243]: DEBUG oslo_concurrency.lockutils [None req-03053e90-e0e5-4e59-9e92-a8c621078633 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "5f1c246d-2732-4da4-a00b-b92e2990aa67" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.272s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.118962] env[61243]: DEBUG oslo_vmware.api [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339419, 'name': PowerOnVM_Task, 'duration_secs': 0.518394} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.119420] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1050.119566] env[61243]: INFO nova.compute.manager [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Took 6.85 seconds to spawn the instance on the hypervisor. [ 1050.119770] env[61243]: DEBUG nova.compute.manager [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1050.120615] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9803b77d-3f87-491a-b960-870f532bce6a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.268741] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "55f24903-b0b4-4124-afde-f31b2680a53a" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.397457] env[61243]: INFO nova.compute.manager [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Rebuilding instance [ 1050.418580] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ee134bd2-d4d1-4787-9b75-5f1be4b5ba66 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.290s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.419426] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.150s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.450577] env[61243]: DEBUG nova.compute.manager [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1050.451644] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98de6c3-cee0-4dc3-828f-17f69a4d16e0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.483352] env[61243]: DEBUG nova.compute.manager [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Stashing vm_state: active {{(pid=61243) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1050.637533] env[61243]: INFO nova.compute.manager [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Took 12.52 seconds to build instance. [ 1050.925337] env[61243]: INFO nova.compute.manager [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Detaching volume 21807d69-846e-400f-b7f6-1d044e77e6c7 [ 1050.967295] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1050.967295] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc9a69e6-0f3f-4268-a3cc-c7860f02fed4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.969785] env[61243]: INFO nova.virt.block_device [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Attempting to driver detach volume 21807d69-846e-400f-b7f6-1d044e77e6c7 from mountpoint /dev/sdb [ 1050.970657] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Volume detach. Driver type: vmdk {{(pid=61243) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1050.970657] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285776', 'volume_id': '21807d69-846e-400f-b7f6-1d044e77e6c7', 'name': 'volume-21807d69-846e-400f-b7f6-1d044e77e6c7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '55f24903-b0b4-4124-afde-f31b2680a53a', 'attached_at': '', 'detached_at': '', 'volume_id': '21807d69-846e-400f-b7f6-1d044e77e6c7', 'serial': '21807d69-846e-400f-b7f6-1d044e77e6c7'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1050.971061] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1045bc74-a93e-4ae1-acc5-1e86ceb561c5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.023315] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ea0689-e176-411a-8a72-238c734ac4d9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.029077] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1051.029077] env[61243]: value = "task-1339420" [ 1051.029077] env[61243]: _type = "Task" [ 1051.029077] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.034545] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Didn't find any instances for network info cache update. {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1051.035875] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.036620] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "3fabef9e-13ff-41b6-a4eb-ee65b66c37fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.036999] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "3fabef9e-13ff-41b6-a4eb-ee65b66c37fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.038740] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "26ca8850-4785-4c02-9dd6-3788d0164510" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.039046] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "26ca8850-4785-4c02-9dd6-3788d0164510" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.040242] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1051.044478] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1051.046376] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1051.048318] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1051.052049] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd4e6c8-e823-48df-a20e-6121fde99cc5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.061656] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1051.062316] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "fbf413f0-389c-4924-bf5d-400f95c89a84" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.062577] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "fbf413f0-389c-4924-bf5d-400f95c89a84" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.064336] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339420, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.064950] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1051.065120] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61243) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1051.066191] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager.update_available_resource {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1051.094980] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80616a9f-95f4-4d13-af32-24095643b894 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.111508] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] The volume has not been displaced from its original location: [datastore2] volume-21807d69-846e-400f-b7f6-1d044e77e6c7/volume-21807d69-846e-400f-b7f6-1d044e77e6c7.vmdk. No consolidation needed. {{(pid=61243) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1051.117387] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Reconfiguring VM instance instance-0000005b to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1051.121057] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9093beca-dd0d-49d2-bbff-f99f350e0fba {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.140157] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2f2cfc32-1727-4300-acd4-94f886e4732b tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.027s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.141911] env[61243]: DEBUG oslo_vmware.api [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 1051.141911] env[61243]: value = "task-1339421" [ 1051.141911] env[61243]: _type = "Task" [ 1051.141911] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.150840] env[61243]: DEBUG oslo_vmware.api [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339421, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.200540] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d99fd2b-72a1-4533-b098-ff62129e4a65 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.208183] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8994cd37-c92b-4990-9985-3f091bf49f65 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.241513] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0630d7-0aba-4fa7-96c7-de0fdb46df12 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.249973] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a773a888-d765-46a3-b22a-f1c9f8409c85 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.265395] env[61243]: DEBUG nova.compute.provider_tree [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.542315] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1051.546024] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1051.557228] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339420, 'name': PowerOffVM_Task, 'duration_secs': 0.277278} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.557228] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1051.565893] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1051.569530] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.630362] env[61243]: INFO nova.compute.manager [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Detaching volume d494ae0c-78b7-4745-85ac-83f826afe53a [ 1051.651954] env[61243]: DEBUG oslo_vmware.api [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339421, 'name': ReconfigVM_Task, 'duration_secs': 0.261985} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.652405] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Reconfigured VM instance instance-0000005b to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1051.657139] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ff46484-9013-43d4-bc71-0e4dbd571981 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.667705] env[61243]: INFO nova.virt.block_device [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Attempting to driver detach volume d494ae0c-78b7-4745-85ac-83f826afe53a from mountpoint /dev/sdb [ 1051.667954] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Volume detach. Driver type: vmdk {{(pid=61243) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1051.668219] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285780', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'name': 'volume-d494ae0c-78b7-4745-85ac-83f826afe53a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5f1c246d-2732-4da4-a00b-b92e2990aa67', 'attached_at': '', 'detached_at': '', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'serial': 'd494ae0c-78b7-4745-85ac-83f826afe53a'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1051.668991] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c50e701-3f40-48a9-b26d-964e12f04bfb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.691181] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98f9ac0d-09b6-4d3e-8120-f643335ebd19 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.694056] env[61243]: DEBUG oslo_vmware.api [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 1051.694056] env[61243]: value = "task-1339422" [ 1051.694056] env[61243]: _type = "Task" [ 1051.694056] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.701727] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8216720-a0fa-48a7-a488-8e20d0da88a4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.707906] env[61243]: DEBUG oslo_vmware.api [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.730537] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f5047a-0837-422f-9bdd-98dd47eb1169 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.748021] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] The volume has not been displaced from its original location: [datastore1] volume-d494ae0c-78b7-4745-85ac-83f826afe53a/volume-d494ae0c-78b7-4745-85ac-83f826afe53a.vmdk. No consolidation needed. {{(pid=61243) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1051.750756] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Reconfiguring VM instance instance-0000005e to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1051.751078] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8916bd7-3d59-4b33-a320-793064759465 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.769364] env[61243]: DEBUG nova.scheduler.client.report [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1051.772588] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1051.772588] env[61243]: value = "task-1339423" [ 1051.772588] env[61243]: _type = "Task" [ 1051.772588] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.781684] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339423, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.912027] env[61243]: DEBUG nova.compute.manager [req-d83a5c1a-a6d3-4d51-8ea1-8a04bef5c6e7 req-bc63e81e-b748-4b54-8779-d06b3d7f580e service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Received event network-changed-de6bbdd5-c1ee-40d7-a91c-1938f53863a7 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1051.912254] env[61243]: DEBUG nova.compute.manager [req-d83a5c1a-a6d3-4d51-8ea1-8a04bef5c6e7 req-bc63e81e-b748-4b54-8779-d06b3d7f580e service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Refreshing instance network info cache due to event network-changed-de6bbdd5-c1ee-40d7-a91c-1938f53863a7. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1051.912480] env[61243]: DEBUG oslo_concurrency.lockutils [req-d83a5c1a-a6d3-4d51-8ea1-8a04bef5c6e7 req-bc63e81e-b748-4b54-8779-d06b3d7f580e service nova] Acquiring lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.912629] env[61243]: DEBUG oslo_concurrency.lockutils [req-d83a5c1a-a6d3-4d51-8ea1-8a04bef5c6e7 req-bc63e81e-b748-4b54-8779-d06b3d7f580e service nova] Acquired lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.912940] env[61243]: DEBUG nova.network.neutron [req-d83a5c1a-a6d3-4d51-8ea1-8a04bef5c6e7 req-bc63e81e-b748-4b54-8779-d06b3d7f580e service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Refreshing network info cache for port de6bbdd5-c1ee-40d7-a91c-1938f53863a7 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1052.073663] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.077775] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.084909] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.204783] env[61243]: DEBUG oslo_vmware.api [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339422, 'name': ReconfigVM_Task, 'duration_secs': 0.174554} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.205146] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285776', 'volume_id': '21807d69-846e-400f-b7f6-1d044e77e6c7', 'name': 'volume-21807d69-846e-400f-b7f6-1d044e77e6c7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '55f24903-b0b4-4124-afde-f31b2680a53a', 'attached_at': '', 'detached_at': '', 'volume_id': '21807d69-846e-400f-b7f6-1d044e77e6c7', 'serial': '21807d69-846e-400f-b7f6-1d044e77e6c7'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1052.274816] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.275380] env[61243]: DEBUG nova.compute.manager [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1052.278351] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.243s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.289748] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339423, 'name': ReconfigVM_Task, 'duration_secs': 0.475354} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.290077] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Reconfigured VM instance instance-0000005e to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1052.295055] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd97f2e6-a5e0-4f38-adea-75e3adfe7243 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.311178] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1052.311178] env[61243]: value = "task-1339424" [ 1052.311178] env[61243]: _type = "Task" [ 1052.311178] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.320318] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339424, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.752680] env[61243]: DEBUG nova.objects.instance [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lazy-loading 'flavor' on Instance uuid 55f24903-b0b4-4124-afde-f31b2680a53a {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1052.774749] env[61243]: DEBUG nova.network.neutron [req-d83a5c1a-a6d3-4d51-8ea1-8a04bef5c6e7 req-bc63e81e-b748-4b54-8779-d06b3d7f580e service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Updated VIF entry in instance network info cache for port de6bbdd5-c1ee-40d7-a91c-1938f53863a7. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1052.775141] env[61243]: DEBUG nova.network.neutron [req-d83a5c1a-a6d3-4d51-8ea1-8a04bef5c6e7 req-bc63e81e-b748-4b54-8779-d06b3d7f580e service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Updating instance_info_cache with network_info: [{"id": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "address": "fa:16:3e:43:9a:6f", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde6bbdd5-c1", "ovs_interfaceid": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.784843] env[61243]: DEBUG nova.compute.utils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1052.787842] env[61243]: INFO nova.compute.claims [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1052.791637] env[61243]: DEBUG nova.compute.manager [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1052.791944] env[61243]: DEBUG nova.network.neutron [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1052.822510] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339424, 'name': ReconfigVM_Task, 'duration_secs': 0.163237} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.822829] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285780', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'name': 'volume-d494ae0c-78b7-4745-85ac-83f826afe53a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5f1c246d-2732-4da4-a00b-b92e2990aa67', 'attached_at': '', 'detached_at': '', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'serial': 'd494ae0c-78b7-4745-85ac-83f826afe53a'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1052.866053] env[61243]: DEBUG nova.policy [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06e247d66c8d43218302a35ead439405', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e641e72e4d2048218ef403456dbb2b82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1053.277793] env[61243]: DEBUG oslo_concurrency.lockutils [req-d83a5c1a-a6d3-4d51-8ea1-8a04bef5c6e7 req-bc63e81e-b748-4b54-8779-d06b3d7f580e service nova] Releasing lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.292516] env[61243]: DEBUG nova.compute.manager [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1053.297826] env[61243]: INFO nova.compute.resource_tracker [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updating resource usage from migration 0fcac3af-14cd-43cd-99d0-3e6f3647002e [ 1053.528489] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b377aab4-8064-4a7d-8f39-4e79e568a76f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.536706] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e21aa36-ab82-4740-b3ae-2d47372d38db {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.540746] env[61243]: DEBUG nova.network.neutron [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Successfully created port: 04ff9536-8994-48d8-aa25-6d4d977da0e6 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1053.568623] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd39c49-f795-41d3-ac8a-7f00f3ce78cc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.576156] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b9c7b4f-bb8c-4d5d-bd75-0ee3d702b2a8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.593549] env[61243]: DEBUG nova.compute.provider_tree [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.693815] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "55f24903-b0b4-4124-afde-f31b2680a53a" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.766456] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f89d3f16-afcf-4096-aa65-5542bac2c707 tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.348s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.767598] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.074s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.891962] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1053.892255] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-777d455f-9b31-41de-a630-9bd969e91956 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.899980] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1053.899980] env[61243]: value = "task-1339425" [ 1053.899980] env[61243]: _type = "Task" [ 1053.899980] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.907673] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339425, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.941268] env[61243]: DEBUG nova.compute.manager [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Received event network-changed-5d2d1a82-6eda-413f-9581-9fa9107be4bd {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1053.941268] env[61243]: DEBUG nova.compute.manager [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Refreshing instance network info cache due to event network-changed-5d2d1a82-6eda-413f-9581-9fa9107be4bd. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1053.941786] env[61243]: DEBUG oslo_concurrency.lockutils [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] Acquiring lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.941786] env[61243]: DEBUG oslo_concurrency.lockutils [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] Acquired lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.941786] env[61243]: DEBUG nova.network.neutron [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Refreshing network info cache for port 5d2d1a82-6eda-413f-9581-9fa9107be4bd {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1054.096920] env[61243]: DEBUG nova.scheduler.client.report [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.271911] env[61243]: INFO nova.compute.manager [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Detaching volume 9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b [ 1054.300847] env[61243]: DEBUG nova.compute.manager [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1054.304384] env[61243]: INFO nova.virt.block_device [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Attempting to driver detach volume 9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b from mountpoint /dev/sdc [ 1054.304630] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Volume detach. Driver type: vmdk {{(pid=61243) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1054.304826] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285781', 'volume_id': '9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b', 'name': 'volume-9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '55f24903-b0b4-4124-afde-f31b2680a53a', 'attached_at': '', 'detached_at': '', 'volume_id': '9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b', 'serial': '9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1054.305746] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aea0f2f-8da8-46cf-9a5a-ec0edc3c6f90 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.311341] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "dca3301d-ef3e-48ef-920c-866b4a086ea6" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.311568] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "dca3301d-ef3e-48ef-920c-866b4a086ea6" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.331623] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e28201d-c3d3-463b-b0cf-9071c0298b32 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.342271] env[61243]: DEBUG nova.virt.hardware [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1054.342529] env[61243]: DEBUG nova.virt.hardware [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1054.342693] env[61243]: DEBUG nova.virt.hardware [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1054.342879] env[61243]: DEBUG nova.virt.hardware [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1054.343044] env[61243]: DEBUG nova.virt.hardware [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1054.343207] env[61243]: DEBUG nova.virt.hardware [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1054.343450] env[61243]: DEBUG nova.virt.hardware [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1054.343630] env[61243]: DEBUG nova.virt.hardware [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1054.343805] env[61243]: DEBUG nova.virt.hardware [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1054.343973] env[61243]: DEBUG nova.virt.hardware [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1054.344167] env[61243]: DEBUG nova.virt.hardware [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1054.344939] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0eb7be-1f8c-40c2-83ca-a5fdd782d696 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.348028] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eae40ab-7ba1-40ba-ad45-f728fd22fa54 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.372558] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbeefcb1-524b-4a0d-a9ea-8449945596ac {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.377836] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e75a57-9647-4eb5-80c4-69b173e082cd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.401649] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] The volume has not been displaced from its original location: [datastore2] volume-9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b/volume-9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b.vmdk. No consolidation needed. {{(pid=61243) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1054.406899] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Reconfiguring VM instance instance-0000005b to detach disk 2002 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1054.410150] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c57a1a17-a41d-4b9e-8bed-4581f01d5a19 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.430205] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] VM already powered off {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1054.430419] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Volume detach. Driver type: vmdk {{(pid=61243) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1054.430624] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285780', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'name': 'volume-d494ae0c-78b7-4745-85ac-83f826afe53a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5f1c246d-2732-4da4-a00b-b92e2990aa67', 'attached_at': '', 'detached_at': '', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'serial': 'd494ae0c-78b7-4745-85ac-83f826afe53a'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1054.430960] env[61243]: DEBUG oslo_vmware.api [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 1054.430960] env[61243]: value = "task-1339426" [ 1054.430960] env[61243]: _type = "Task" [ 1054.430960] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.431722] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfbe9d4c-e74b-4cdc-8c1f-e3da179f69ca {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.458239] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dadd4bf4-db85-4379-b934-ae076f7043aa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.461010] env[61243]: DEBUG oslo_vmware.api [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339426, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.468396] env[61243]: WARNING nova.virt.vmwareapi.driver [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1054.468396] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1054.468396] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-978e5a49-ff80-4eac-afc4-e0fbd9a2cc1f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.474342] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1054.477070] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b819a1a3-7e13-4aaf-afc8-840534d405e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.564470] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1054.564812] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1054.565088] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleting the datastore file [datastore1] 5f1c246d-2732-4da4-a00b-b92e2990aa67 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1054.565381] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-466ba44b-b146-4bee-be43-b6794699cf00 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.573617] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1054.573617] env[61243]: value = "task-1339428" [ 1054.573617] env[61243]: _type = "Task" [ 1054.573617] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.587391] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339428, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.608584] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.330s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.608887] env[61243]: INFO nova.compute.manager [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Migrating [ 1054.617295] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.048s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.618102] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.618102] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61243) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1054.618102] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.545s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.619579] env[61243]: INFO nova.compute.claims [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1054.628905] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef01f5f1-756c-4fc3-b7fc-b1fb05640fee {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.643920] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3580f1-b829-4aba-8fd6-ad1fba5c80c6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.660921] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef770108-0470-4051-9927-30a8d63ebac7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.672472] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e548e0-fa5c-4cf4-af8c-a5d4c6e242b7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.709730] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180571MB free_disk=154GB free_vcpus=48 pci_devices=None {{(pid=61243) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1054.709899] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.717035] env[61243]: DEBUG nova.network.neutron [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Updated VIF entry in instance network info cache for port 5d2d1a82-6eda-413f-9581-9fa9107be4bd. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1054.717421] env[61243]: DEBUG nova.network.neutron [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Updating instance_info_cache with network_info: [{"id": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "address": "fa:16:3e:a4:7d:cc", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d2d1a82-6e", "ovs_interfaceid": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.831988] env[61243]: DEBUG nova.compute.utils [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1054.945816] env[61243]: DEBUG oslo_vmware.api [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339426, 'name': ReconfigVM_Task, 'duration_secs': 0.258974} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.946143] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Reconfigured VM instance instance-0000005b to detach disk 2002 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1054.950889] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21ba6ac5-16c7-49ad-a1a7-a75e037c7aae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.970672] env[61243]: DEBUG oslo_vmware.api [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 1054.970672] env[61243]: value = "task-1339429" [ 1054.970672] env[61243]: _type = "Task" [ 1054.970672] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.978072] env[61243]: DEBUG oslo_vmware.api [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339429, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.083535] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339428, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172468} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.083958] env[61243]: DEBUG nova.network.neutron [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Successfully updated port: 04ff9536-8994-48d8-aa25-6d4d977da0e6 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1055.085178] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1055.085382] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1055.085567] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1055.138677] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "refresh_cache-bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.138916] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "refresh_cache-bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.139128] env[61243]: DEBUG nova.network.neutron [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1055.220311] env[61243]: DEBUG oslo_concurrency.lockutils [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] Releasing lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.220577] env[61243]: DEBUG nova.compute.manager [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Received event network-changed-5d2d1a82-6eda-413f-9581-9fa9107be4bd {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.220751] env[61243]: DEBUG nova.compute.manager [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Refreshing instance network info cache due to event network-changed-5d2d1a82-6eda-413f-9581-9fa9107be4bd. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1055.220962] env[61243]: DEBUG oslo_concurrency.lockutils [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] Acquiring lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.221125] env[61243]: DEBUG oslo_concurrency.lockutils [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] Acquired lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.221352] env[61243]: DEBUG nova.network.neutron [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Refreshing network info cache for port 5d2d1a82-6eda-413f-9581-9fa9107be4bd {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1055.335425] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "dca3301d-ef3e-48ef-920c-866b4a086ea6" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.024s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.480755] env[61243]: DEBUG oslo_vmware.api [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339429, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.589131] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "refresh_cache-df141b12-1d34-4a47-b6bb-15cb75d266c3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.589364] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "refresh_cache-df141b12-1d34-4a47-b6bb-15cb75d266c3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.589578] env[61243]: DEBUG nova.network.neutron [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1055.592567] env[61243]: INFO nova.virt.block_device [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Booting with volume d494ae0c-78b7-4745-85ac-83f826afe53a at /dev/sdb [ 1055.628412] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-044280b8-10b6-4ef3-b0b2-e7ff46c19c9a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.639359] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c90d9e-26c0-474b-a6e0-3af7f5fbd437 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.669694] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-11582d7a-0f32-4af6-9cee-476d4478d0c3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.681481] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e26a54d-5bda-43c4-b25e-41c585ee66ea {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.720192] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-471b3196-fdfe-450b-802e-7c901a8c7496 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.728558] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9155df73-6e05-45ca-b4ac-4da912e5df7c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.741783] env[61243]: DEBUG nova.virt.block_device [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Updating existing volume attachment record: b485ddf7-2023-49df-a7ea-88414c864e08 {{(pid=61243) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1055.869031] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30b412f4-a122-44f6-8fde-bcdd89ef1ae4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.879148] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db1658b-5741-4a0f-a072-6e1cefd4cda8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.918798] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "dca3301d-ef3e-48ef-920c-866b4a086ea6" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.919669] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "dca3301d-ef3e-48ef-920c-866b4a086ea6" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.919669] env[61243]: INFO nova.compute.manager [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Attaching volume bce221b9-8c01-45eb-b91e-16b64854ec9c to /dev/sdb [ 1055.921412] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa441f9-e6e2-4a4c-8389-254677aa80cf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.933709] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc0eb94-5d42-4e4c-9905-e5436967c795 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.947680] env[61243]: DEBUG nova.compute.provider_tree [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.962384] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4abe3362-fe4a-4c9a-a1fa-ae5cff28c3f3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.976685] env[61243]: DEBUG nova.compute.manager [req-6937cefa-8bee-49b5-8109-f9c1400cbce1 req-f053e0d8-e045-4ad7-90d0-605793cdc4e1 service nova] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Received event network-vif-plugged-04ff9536-8994-48d8-aa25-6d4d977da0e6 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.977054] env[61243]: DEBUG oslo_concurrency.lockutils [req-6937cefa-8bee-49b5-8109-f9c1400cbce1 req-f053e0d8-e045-4ad7-90d0-605793cdc4e1 service nova] Acquiring lock "df141b12-1d34-4a47-b6bb-15cb75d266c3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.977245] env[61243]: DEBUG oslo_concurrency.lockutils [req-6937cefa-8bee-49b5-8109-f9c1400cbce1 req-f053e0d8-e045-4ad7-90d0-605793cdc4e1 service nova] Lock "df141b12-1d34-4a47-b6bb-15cb75d266c3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.977437] env[61243]: DEBUG oslo_concurrency.lockutils [req-6937cefa-8bee-49b5-8109-f9c1400cbce1 req-f053e0d8-e045-4ad7-90d0-605793cdc4e1 service nova] Lock "df141b12-1d34-4a47-b6bb-15cb75d266c3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.977565] env[61243]: DEBUG nova.compute.manager [req-6937cefa-8bee-49b5-8109-f9c1400cbce1 req-f053e0d8-e045-4ad7-90d0-605793cdc4e1 service nova] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] No waiting events found dispatching network-vif-plugged-04ff9536-8994-48d8-aa25-6d4d977da0e6 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1055.977972] env[61243]: WARNING nova.compute.manager [req-6937cefa-8bee-49b5-8109-f9c1400cbce1 req-f053e0d8-e045-4ad7-90d0-605793cdc4e1 service nova] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Received unexpected event network-vif-plugged-04ff9536-8994-48d8-aa25-6d4d977da0e6 for instance with vm_state building and task_state spawning. [ 1055.977972] env[61243]: DEBUG nova.compute.manager [req-6937cefa-8bee-49b5-8109-f9c1400cbce1 req-f053e0d8-e045-4ad7-90d0-605793cdc4e1 service nova] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Received event network-changed-04ff9536-8994-48d8-aa25-6d4d977da0e6 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.978079] env[61243]: DEBUG nova.compute.manager [req-6937cefa-8bee-49b5-8109-f9c1400cbce1 req-f053e0d8-e045-4ad7-90d0-605793cdc4e1 service nova] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Refreshing instance network info cache due to event network-changed-04ff9536-8994-48d8-aa25-6d4d977da0e6. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1055.978233] env[61243]: DEBUG oslo_concurrency.lockutils [req-6937cefa-8bee-49b5-8109-f9c1400cbce1 req-f053e0d8-e045-4ad7-90d0-605793cdc4e1 service nova] Acquiring lock "refresh_cache-df141b12-1d34-4a47-b6bb-15cb75d266c3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.979696] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f610871-a693-46e3-a366-864b7632457c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.988507] env[61243]: DEBUG oslo_vmware.api [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339429, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.995622] env[61243]: DEBUG nova.network.neutron [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updating instance_info_cache with network_info: [{"id": "a968f6e7-b39a-4dd8-abf8-bf0919565a72", "address": "fa:16:3e:80:14:77", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa968f6e7-b3", "ovs_interfaceid": "a968f6e7-b39a-4dd8-abf8-bf0919565a72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.000405] env[61243]: DEBUG nova.virt.block_device [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Updating existing volume attachment record: a508ce16-ab5f-4b88-9805-3914c71ff90b {{(pid=61243) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1056.052558] env[61243]: DEBUG nova.network.neutron [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Updated VIF entry in instance network info cache for port 5d2d1a82-6eda-413f-9581-9fa9107be4bd. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1056.052993] env[61243]: DEBUG nova.network.neutron [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Updating instance_info_cache with network_info: [{"id": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "address": "fa:16:3e:a4:7d:cc", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d2d1a82-6e", "ovs_interfaceid": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.128470] env[61243]: DEBUG nova.network.neutron [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1056.261756] env[61243]: DEBUG nova.network.neutron [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Updating instance_info_cache with network_info: [{"id": "04ff9536-8994-48d8-aa25-6d4d977da0e6", "address": "fa:16:3e:d3:3a:ee", "network": {"id": "bce1aecb-e043-4f68-99e2-c6b03c4bb866", "bridge": "br-int", "label": "tempest-ServersTestJSON-652904947-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e641e72e4d2048218ef403456dbb2b82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04ff9536-89", "ovs_interfaceid": "04ff9536-8994-48d8-aa25-6d4d977da0e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.451204] env[61243]: DEBUG nova.scheduler.client.report [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1056.479409] env[61243]: DEBUG oslo_vmware.api [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339429, 'name': ReconfigVM_Task, 'duration_secs': 1.147476} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.479777] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285781', 'volume_id': '9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b', 'name': 'volume-9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '55f24903-b0b4-4124-afde-f31b2680a53a', 'attached_at': '', 'detached_at': '', 'volume_id': '9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b', 'serial': '9ccead0d-7cbe-4eec-bf14-cdaca91e9b6b'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1056.504263] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "refresh_cache-bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.557195] env[61243]: DEBUG oslo_concurrency.lockutils [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] Releasing lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.557498] env[61243]: DEBUG nova.compute.manager [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Received event network-changed-de6bbdd5-c1ee-40d7-a91c-1938f53863a7 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1056.557674] env[61243]: DEBUG nova.compute.manager [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Refreshing instance network info cache due to event network-changed-de6bbdd5-c1ee-40d7-a91c-1938f53863a7. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1056.557896] env[61243]: DEBUG oslo_concurrency.lockutils [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] Acquiring lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.558054] env[61243]: DEBUG oslo_concurrency.lockutils [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] Acquired lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.558228] env[61243]: DEBUG nova.network.neutron [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Refreshing network info cache for port de6bbdd5-c1ee-40d7-a91c-1938f53863a7 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1056.764620] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "refresh_cache-df141b12-1d34-4a47-b6bb-15cb75d266c3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.764871] env[61243]: DEBUG nova.compute.manager [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Instance network_info: |[{"id": "04ff9536-8994-48d8-aa25-6d4d977da0e6", "address": "fa:16:3e:d3:3a:ee", "network": {"id": "bce1aecb-e043-4f68-99e2-c6b03c4bb866", "bridge": "br-int", "label": "tempest-ServersTestJSON-652904947-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e641e72e4d2048218ef403456dbb2b82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04ff9536-89", "ovs_interfaceid": "04ff9536-8994-48d8-aa25-6d4d977da0e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1056.765225] env[61243]: DEBUG oslo_concurrency.lockutils [req-6937cefa-8bee-49b5-8109-f9c1400cbce1 req-f053e0d8-e045-4ad7-90d0-605793cdc4e1 service nova] Acquired lock "refresh_cache-df141b12-1d34-4a47-b6bb-15cb75d266c3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.765421] env[61243]: DEBUG nova.network.neutron [req-6937cefa-8bee-49b5-8109-f9c1400cbce1 req-f053e0d8-e045-4ad7-90d0-605793cdc4e1 service nova] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Refreshing network info cache for port 04ff9536-8994-48d8-aa25-6d4d977da0e6 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1056.766674] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d3:3a:ee', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '90878b7b-ddb7-4f47-892b-d6e06f73475f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '04ff9536-8994-48d8-aa25-6d4d977da0e6', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1056.774806] env[61243]: DEBUG oslo.service.loopingcall [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1056.775889] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1056.776144] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5a129772-8bea-4a10-90c6-571375627986 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.798930] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1056.798930] env[61243]: value = "task-1339431" [ 1056.798930] env[61243]: _type = "Task" [ 1056.798930] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.806775] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339431, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.961435] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.339s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.961435] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1056.961894] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.882s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.963325] env[61243]: INFO nova.compute.claims [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1057.027711] env[61243]: DEBUG nova.objects.instance [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lazy-loading 'flavor' on Instance uuid 55f24903-b0b4-4124-afde-f31b2680a53a {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1057.308870] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339431, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.320569] env[61243]: DEBUG nova.network.neutron [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Updated VIF entry in instance network info cache for port de6bbdd5-c1ee-40d7-a91c-1938f53863a7. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1057.321238] env[61243]: DEBUG nova.network.neutron [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Updating instance_info_cache with network_info: [{"id": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "address": "fa:16:3e:43:9a:6f", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde6bbdd5-c1", "ovs_interfaceid": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.468029] env[61243]: DEBUG nova.compute.utils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1057.473935] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1057.474160] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1057.506251] env[61243]: DEBUG nova.network.neutron [req-6937cefa-8bee-49b5-8109-f9c1400cbce1 req-f053e0d8-e045-4ad7-90d0-605793cdc4e1 service nova] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Updated VIF entry in instance network info cache for port 04ff9536-8994-48d8-aa25-6d4d977da0e6. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1057.506714] env[61243]: DEBUG nova.network.neutron [req-6937cefa-8bee-49b5-8109-f9c1400cbce1 req-f053e0d8-e045-4ad7-90d0-605793cdc4e1 service nova] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Updating instance_info_cache with network_info: [{"id": "04ff9536-8994-48d8-aa25-6d4d977da0e6", "address": "fa:16:3e:d3:3a:ee", "network": {"id": "bce1aecb-e043-4f68-99e2-c6b03c4bb866", "bridge": "br-int", "label": "tempest-ServersTestJSON-652904947-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e641e72e4d2048218ef403456dbb2b82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04ff9536-89", "ovs_interfaceid": "04ff9536-8994-48d8-aa25-6d4d977da0e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.527557] env[61243]: DEBUG nova.policy [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ceff73e011724c0eaaa89b1012cabdfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78bce9f020ee429e9902f3fad3b62a3c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1057.783029] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Successfully created port: f8c67797-fbfc-41d3-b441-9c5274d8c188 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1057.810787] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339431, 'name': CreateVM_Task, 'duration_secs': 0.52986} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.810986] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1057.811686] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.811869] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.812246] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1057.812486] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0af61562-2b63-4272-9a1e-bfa74db1616e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.817296] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1057.817296] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]526dfb11-98f0-8acc-1d3f-597e2ff6ee80" [ 1057.817296] env[61243]: _type = "Task" [ 1057.817296] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.825283] env[61243]: DEBUG oslo_concurrency.lockutils [req-e4ca9ea0-34af-471e-9d8d-7fdde05d8e30 req-f50b672b-a4af-401c-8ba3-ced9c738e3aa service nova] Releasing lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.825685] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]526dfb11-98f0-8acc-1d3f-597e2ff6ee80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.975045] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1058.009392] env[61243]: DEBUG oslo_concurrency.lockutils [req-6937cefa-8bee-49b5-8109-f9c1400cbce1 req-f053e0d8-e045-4ad7-90d0-605793cdc4e1 service nova] Releasing lock "refresh_cache-df141b12-1d34-4a47-b6bb-15cb75d266c3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.021614] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91791537-c6af-425e-a05e-844caeb56c69 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.041216] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0ee8708a-0366-4dcf-a66f-80aa38e01e9c tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.273s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.042104] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updating instance 'bb7c810f-91d1-43eb-b0cc-fb4c686b1f29' progress to 0 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1058.087039] env[61243]: DEBUG nova.virt.hardware [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1058.087319] env[61243]: DEBUG nova.virt.hardware [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1058.087485] env[61243]: DEBUG nova.virt.hardware [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1058.087674] env[61243]: DEBUG nova.virt.hardware [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1058.087828] env[61243]: DEBUG nova.virt.hardware [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1058.087983] env[61243]: DEBUG nova.virt.hardware [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1058.088222] env[61243]: DEBUG nova.virt.hardware [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1058.088398] env[61243]: DEBUG nova.virt.hardware [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1058.088554] env[61243]: DEBUG nova.virt.hardware [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1058.088721] env[61243]: DEBUG nova.virt.hardware [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1058.088898] env[61243]: DEBUG nova.virt.hardware [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1058.091951] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25bf5bfe-c267-474d-b1c4-1e07459b1def {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.101578] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e0c878-33bc-41ff-b4ea-15f879d924a2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.119149] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:9e:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ff3ecd2f-0b10-4faf-a512-fd7a20c28df1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6fddda84-0767-4709-ac83-fc4e6060da3c', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1058.126999] env[61243]: DEBUG oslo.service.loopingcall [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1058.129819] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1058.130249] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bf36cbd1-8f61-4d19-9ba5-34da507c28e4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.152314] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1058.152314] env[61243]: value = "task-1339432" [ 1058.152314] env[61243]: _type = "Task" [ 1058.152314] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.159759] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339432, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.218460] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47584a91-a059-4e3b-8418-4dd00a6feb33 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.227666] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9afecf02-9e09-4d3e-99fe-f26d8e04bc65 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.256785] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d92fd63-120b-494c-9973-3109d1c10f3f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.264274] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d719f900-a6d7-47ae-a946-ace408c0f83a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.277246] env[61243]: DEBUG nova.compute.provider_tree [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.326994] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]526dfb11-98f0-8acc-1d3f-597e2ff6ee80, 'name': SearchDatastore_Task, 'duration_secs': 0.010366} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.327262] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.327506] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1058.327747] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.327900] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.328100] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1058.328364] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c5da337a-7fb4-4ad7-860c-14b71e6afa26 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.337068] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1058.337068] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1058.337763] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b37eca5-8ba3-4fb7-9fd8-308e424208c3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.342607] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1058.342607] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52fe1e16-0483-528a-bfb9-f9a17bbd1e65" [ 1058.342607] env[61243]: _type = "Task" [ 1058.342607] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.350192] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52fe1e16-0483-528a-bfb9-f9a17bbd1e65, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.550621] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1058.550968] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9faec487-658d-4cce-9a59-c8d4dc0cf887 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.557495] env[61243]: DEBUG oslo_vmware.api [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1058.557495] env[61243]: value = "task-1339433" [ 1058.557495] env[61243]: _type = "Task" [ 1058.557495] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.565993] env[61243]: DEBUG oslo_vmware.api [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339433, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.663754] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339432, 'name': CreateVM_Task, 'duration_secs': 0.34798} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.664665] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1058.665404] env[61243]: DEBUG oslo_concurrency.lockutils [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.665656] env[61243]: DEBUG oslo_concurrency.lockutils [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.665956] env[61243]: DEBUG oslo_concurrency.lockutils [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1058.666536] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7446b61-6d45-4a3a-9aac-81d265ec3d49 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.674483] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1058.674483] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5297d94a-8546-6441-bdc8-bd8a851bea39" [ 1058.674483] env[61243]: _type = "Task" [ 1058.674483] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.682948] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5297d94a-8546-6441-bdc8-bd8a851bea39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.780181] env[61243]: DEBUG nova.scheduler.client.report [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1058.852574] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52fe1e16-0483-528a-bfb9-f9a17bbd1e65, 'name': SearchDatastore_Task, 'duration_secs': 0.008677} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.853102] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab4e9d40-9b37-498a-a313-691a697b54a8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.858448] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1058.858448] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d83492-aa18-639d-8c76-8e78c89587f9" [ 1058.858448] env[61243]: _type = "Task" [ 1058.858448] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.866223] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d83492-aa18-639d-8c76-8e78c89587f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.987187] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1059.012744] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1059.013008] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1059.013211] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1059.013436] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1059.013591] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1059.013746] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1059.013957] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1059.014141] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1059.014329] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1059.014480] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1059.014659] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1059.016037] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c468740-06b0-4848-9b98-40309419d18c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.023670] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ba1352-1701-45f4-a0b5-e78bb68378ae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.066824] env[61243]: DEBUG oslo_vmware.api [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339433, 'name': PowerOffVM_Task, 'duration_secs': 0.182881} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.067146] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1059.067344] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updating instance 'bb7c810f-91d1-43eb-b0cc-fb4c686b1f29' progress to 17 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1059.154734] env[61243]: DEBUG oslo_concurrency.lockutils [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "55f24903-b0b4-4124-afde-f31b2680a53a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.155106] env[61243]: DEBUG oslo_concurrency.lockutils [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.155256] env[61243]: DEBUG oslo_concurrency.lockutils [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "55f24903-b0b4-4124-afde-f31b2680a53a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.155467] env[61243]: DEBUG oslo_concurrency.lockutils [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.155679] env[61243]: DEBUG oslo_concurrency.lockutils [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.157894] env[61243]: INFO nova.compute.manager [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Terminating instance [ 1059.159771] env[61243]: DEBUG nova.compute.manager [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1059.159971] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1059.160820] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-625fe24a-09f8-4ba5-94f0-ca496d14a840 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.168821] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1059.169026] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-547cc02d-542e-45ab-8c77-1ce6eb544c06 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.174958] env[61243]: DEBUG oslo_vmware.api [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 1059.174958] env[61243]: value = "task-1339434" [ 1059.174958] env[61243]: _type = "Task" [ 1059.174958] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.184205] env[61243]: DEBUG nova.compute.manager [req-b602a596-a31e-416c-8bfe-47f6cae2398f req-ec60ca85-c255-460e-8bac-4d341dc0341b service nova] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Received event network-vif-plugged-f8c67797-fbfc-41d3-b441-9c5274d8c188 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.184535] env[61243]: DEBUG oslo_concurrency.lockutils [req-b602a596-a31e-416c-8bfe-47f6cae2398f req-ec60ca85-c255-460e-8bac-4d341dc0341b service nova] Acquiring lock "3fabef9e-13ff-41b6-a4eb-ee65b66c37fe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.184695] env[61243]: DEBUG oslo_concurrency.lockutils [req-b602a596-a31e-416c-8bfe-47f6cae2398f req-ec60ca85-c255-460e-8bac-4d341dc0341b service nova] Lock "3fabef9e-13ff-41b6-a4eb-ee65b66c37fe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.184864] env[61243]: DEBUG oslo_concurrency.lockutils [req-b602a596-a31e-416c-8bfe-47f6cae2398f req-ec60ca85-c255-460e-8bac-4d341dc0341b service nova] Lock "3fabef9e-13ff-41b6-a4eb-ee65b66c37fe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.185052] env[61243]: DEBUG nova.compute.manager [req-b602a596-a31e-416c-8bfe-47f6cae2398f req-ec60ca85-c255-460e-8bac-4d341dc0341b service nova] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] No waiting events found dispatching network-vif-plugged-f8c67797-fbfc-41d3-b441-9c5274d8c188 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1059.185229] env[61243]: WARNING nova.compute.manager [req-b602a596-a31e-416c-8bfe-47f6cae2398f req-ec60ca85-c255-460e-8bac-4d341dc0341b service nova] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Received unexpected event network-vif-plugged-f8c67797-fbfc-41d3-b441-9c5274d8c188 for instance with vm_state building and task_state spawning. [ 1059.191910] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5297d94a-8546-6441-bdc8-bd8a851bea39, 'name': SearchDatastore_Task, 'duration_secs': 0.010069} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.192160] env[61243]: DEBUG oslo_vmware.api [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339434, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.192529] env[61243]: DEBUG oslo_concurrency.lockutils [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.192796] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1059.193093] env[61243]: DEBUG oslo_concurrency.lockutils [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.286628] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.326s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.286819] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1059.289352] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.205s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.291570] env[61243]: INFO nova.compute.claims [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1059.370772] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d83492-aa18-639d-8c76-8e78c89587f9, 'name': SearchDatastore_Task, 'duration_secs': 0.009332} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.371112] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.371356] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] df141b12-1d34-4a47-b6bb-15cb75d266c3/df141b12-1d34-4a47-b6bb-15cb75d266c3.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1059.371644] env[61243]: DEBUG oslo_concurrency.lockutils [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.371840] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1059.372103] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84255466-153d-4536-bb9a-1237fa7f009c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.374197] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0bd33d14-e183-4b68-b61b-b7b3c3e24862 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.379929] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1059.379929] env[61243]: value = "task-1339435" [ 1059.379929] env[61243]: _type = "Task" [ 1059.379929] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.389234] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339435, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.399488] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1059.399734] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1059.400507] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a4e0126-2173-4611-abd2-4d5b3325644b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.405762] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1059.405762] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cd7dc1-e426-215d-cdd0-31511d8fbf8f" [ 1059.405762] env[61243]: _type = "Task" [ 1059.405762] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.414752] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cd7dc1-e426-215d-cdd0-31511d8fbf8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.573812] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1059.574102] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1059.574360] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1059.574592] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1059.574769] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1059.574945] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1059.575179] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1059.575400] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1059.575539] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1059.575742] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1059.575991] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1059.582441] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-985a7828-d919-46a9-8cbd-56c075e58ccd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.600482] env[61243]: DEBUG oslo_vmware.api [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1059.600482] env[61243]: value = "task-1339436" [ 1059.600482] env[61243]: _type = "Task" [ 1059.600482] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.609303] env[61243]: DEBUG oslo_vmware.api [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339436, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.685872] env[61243]: DEBUG oslo_vmware.api [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339434, 'name': PowerOffVM_Task, 'duration_secs': 0.208048} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.686205] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1059.686395] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1059.686692] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f0e9fd40-9997-44a9-9da8-d6f226a82d83 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.758588] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1059.758887] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1059.759302] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Deleting the datastore file [datastore2] 55f24903-b0b4-4124-afde-f31b2680a53a {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1059.759752] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-84fb75b7-6513-4e6e-8985-b6fb4c22f266 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.764481] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Successfully updated port: f8c67797-fbfc-41d3-b441-9c5274d8c188 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1059.771988] env[61243]: DEBUG oslo_vmware.api [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for the task: (returnval){ [ 1059.771988] env[61243]: value = "task-1339438" [ 1059.771988] env[61243]: _type = "Task" [ 1059.771988] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.788383] env[61243]: DEBUG oslo_vmware.api [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339438, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.793050] env[61243]: DEBUG nova.compute.manager [req-b1cb5064-3a51-4fdb-a328-2ac9f48bf893 req-2edc3e74-5533-4523-b7b3-31d3b455a677 service nova] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Received event network-changed-f8c67797-fbfc-41d3-b441-9c5274d8c188 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.793050] env[61243]: DEBUG nova.compute.manager [req-b1cb5064-3a51-4fdb-a328-2ac9f48bf893 req-2edc3e74-5533-4523-b7b3-31d3b455a677 service nova] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Refreshing instance network info cache due to event network-changed-f8c67797-fbfc-41d3-b441-9c5274d8c188. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1059.793050] env[61243]: DEBUG oslo_concurrency.lockutils [req-b1cb5064-3a51-4fdb-a328-2ac9f48bf893 req-2edc3e74-5533-4523-b7b3-31d3b455a677 service nova] Acquiring lock "refresh_cache-3fabef9e-13ff-41b6-a4eb-ee65b66c37fe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.793376] env[61243]: DEBUG oslo_concurrency.lockutils [req-b1cb5064-3a51-4fdb-a328-2ac9f48bf893 req-2edc3e74-5533-4523-b7b3-31d3b455a677 service nova] Acquired lock "refresh_cache-3fabef9e-13ff-41b6-a4eb-ee65b66c37fe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.793655] env[61243]: DEBUG nova.network.neutron [req-b1cb5064-3a51-4fdb-a328-2ac9f48bf893 req-2edc3e74-5533-4523-b7b3-31d3b455a677 service nova] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Refreshing network info cache for port f8c67797-fbfc-41d3-b441-9c5274d8c188 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1059.797740] env[61243]: DEBUG nova.compute.utils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1059.805372] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1059.805563] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1059.857113] env[61243]: DEBUG nova.network.neutron [req-b1cb5064-3a51-4fdb-a328-2ac9f48bf893 req-2edc3e74-5533-4523-b7b3-31d3b455a677 service nova] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1059.889368] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339435, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.890841] env[61243]: DEBUG nova.policy [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ceff73e011724c0eaaa89b1012cabdfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78bce9f020ee429e9902f3fad3b62a3c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1059.915124] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cd7dc1-e426-215d-cdd0-31511d8fbf8f, 'name': SearchDatastore_Task, 'duration_secs': 0.00922} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.915905] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66e5740b-c386-4839-9579-f8de028d39fd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.920769] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1059.920769] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cbdc9a-29bd-2b8e-d2ec-fc659ad57a9a" [ 1059.920769] env[61243]: _type = "Task" [ 1059.920769] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.929559] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cbdc9a-29bd-2b8e-d2ec-fc659ad57a9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.984143] env[61243]: DEBUG nova.network.neutron [req-b1cb5064-3a51-4fdb-a328-2ac9f48bf893 req-2edc3e74-5533-4523-b7b3-31d3b455a677 service nova] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.110415] env[61243]: DEBUG oslo_vmware.api [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339436, 'name': ReconfigVM_Task, 'duration_secs': 0.201904} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.112979] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updating instance 'bb7c810f-91d1-43eb-b0cc-fb4c686b1f29' progress to 33 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1060.191944] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Successfully created port: 78eac2ea-b988-4138-a10a-a0fd00a3cdb1 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1060.268392] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "refresh_cache-3fabef9e-13ff-41b6-a4eb-ee65b66c37fe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.282960] env[61243]: DEBUG oslo_vmware.api [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Task: {'id': task-1339438, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191736} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.286413] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1060.286634] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1060.286821] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1060.287015] env[61243]: INFO nova.compute.manager [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1060.287420] env[61243]: DEBUG oslo.service.loopingcall [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1060.287633] env[61243]: DEBUG nova.compute.manager [-] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1060.287734] env[61243]: DEBUG nova.network.neutron [-] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1060.303739] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1060.390945] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339435, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517541} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.395832] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] df141b12-1d34-4a47-b6bb-15cb75d266c3/df141b12-1d34-4a47-b6bb-15cb75d266c3.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1060.396065] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1060.397243] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa39afa6-b071-4046-a17d-76535f61b39b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.403289] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1060.403289] env[61243]: value = "task-1339440" [ 1060.403289] env[61243]: _type = "Task" [ 1060.403289] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.415128] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339440, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.434208] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cbdc9a-29bd-2b8e-d2ec-fc659ad57a9a, 'name': SearchDatastore_Task, 'duration_secs': 0.008083} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.434550] env[61243]: DEBUG oslo_concurrency.lockutils [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.435010] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 5f1c246d-2732-4da4-a00b-b92e2990aa67/5f1c246d-2732-4da4-a00b-b92e2990aa67.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1060.435355] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-419cf1d8-cb0c-4125-a3c4-37f5a97026ba {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.446848] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1060.446848] env[61243]: value = "task-1339441" [ 1060.446848] env[61243]: _type = "Task" [ 1060.446848] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.459765] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339441, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.488950] env[61243]: DEBUG oslo_concurrency.lockutils [req-b1cb5064-3a51-4fdb-a328-2ac9f48bf893 req-2edc3e74-5533-4523-b7b3-31d3b455a677 service nova] Releasing lock "refresh_cache-3fabef9e-13ff-41b6-a4eb-ee65b66c37fe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.490211] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquired lock "refresh_cache-3fabef9e-13ff-41b6-a4eb-ee65b66c37fe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.490211] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1060.544494] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f4143a-1781-450f-907b-0d0603931e1a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.552446] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37e3bae1-e038-418c-bc5e-61abc55b9367 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.585964] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04475dec-b980-451d-961a-95ff7fb50eda {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.594479] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3889b8-2556-4576-b0b2-51853560f086 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.610111] env[61243]: DEBUG nova.compute.provider_tree [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.619021] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1060.619021] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1060.619021] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1060.619786] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1060.619786] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1060.619786] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1060.619786] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1060.619786] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1060.620413] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1060.620413] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1060.620413] env[61243]: DEBUG nova.virt.hardware [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1060.626076] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1060.627055] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5b0744c-a519-40b0-94fb-2824ced205c5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.646536] env[61243]: DEBUG oslo_vmware.api [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1060.646536] env[61243]: value = "task-1339442" [ 1060.646536] env[61243]: _type = "Task" [ 1060.646536] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.656809] env[61243]: DEBUG oslo_vmware.api [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339442, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.913959] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339440, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061873} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.914791] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1060.914946] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e44674e1-430d-4c81-b30e-49a654da6a71 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.938546] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] df141b12-1d34-4a47-b6bb-15cb75d266c3/df141b12-1d34-4a47-b6bb-15cb75d266c3.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1060.939254] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-654763a4-7596-4e9e-8839-721f7ddbca1f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.962417] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339441, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.963757] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1060.963757] env[61243]: value = "task-1339443" [ 1060.963757] env[61243]: _type = "Task" [ 1060.963757] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.972149] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339443, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.024554] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1061.114228] env[61243]: DEBUG nova.scheduler.client.report [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1061.161454] env[61243]: DEBUG oslo_vmware.api [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339442, 'name': ReconfigVM_Task, 'duration_secs': 0.386233} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.161849] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1061.163338] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6621e689-6945-4470-a90c-7336f0eee727 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.186889] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] bb7c810f-91d1-43eb-b0cc-fb4c686b1f29/bb7c810f-91d1-43eb-b0cc-fb4c686b1f29.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1061.187736] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-588d3945-2473-4661-b33a-c5781280748f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.212235] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Updating instance_info_cache with network_info: [{"id": "f8c67797-fbfc-41d3-b441-9c5274d8c188", "address": "fa:16:3e:0a:72:7d", "network": {"id": "2ec6398f-b75e-44f0-a1c8-dffea7f87b26", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-177843661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78bce9f020ee429e9902f3fad3b62a3c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8c67797-fb", "ovs_interfaceid": "f8c67797-fbfc-41d3-b441-9c5274d8c188", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.215597] env[61243]: DEBUG oslo_vmware.api [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1061.215597] env[61243]: value = "task-1339444" [ 1061.215597] env[61243]: _type = "Task" [ 1061.215597] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.226487] env[61243]: DEBUG oslo_vmware.api [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339444, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.314132] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1061.344424] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1061.344817] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1061.345091] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1061.345298] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1061.345493] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1061.345707] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1061.345952] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1061.346141] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1061.346342] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1061.346553] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1061.346832] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1061.347980] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a3faa7-0549-4296-897f-4633b0a69f5b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.356851] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38bcde7d-ce6c-46fd-bf7e-59556ae0a3f8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.391128] env[61243]: DEBUG nova.network.neutron [-] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.463669] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339441, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.924224} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.464473] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 5f1c246d-2732-4da4-a00b-b92e2990aa67/5f1c246d-2732-4da4-a00b-b92e2990aa67.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1061.464473] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1061.464749] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e1683b0d-6df3-40a6-84b0-3c5c9a062db9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.475012] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339443, 'name': ReconfigVM_Task, 'duration_secs': 0.487806} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.476205] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Reconfigured VM instance instance-00000068 to attach disk [datastore1] df141b12-1d34-4a47-b6bb-15cb75d266c3/df141b12-1d34-4a47-b6bb-15cb75d266c3.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1061.476905] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1061.476905] env[61243]: value = "task-1339445" [ 1061.476905] env[61243]: _type = "Task" [ 1061.476905] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.477125] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-021e9b43-51cf-4721-b419-4d81cfe1b386 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.486862] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339445, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.488301] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1061.488301] env[61243]: value = "task-1339446" [ 1061.488301] env[61243]: _type = "Task" [ 1061.488301] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.496994] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339446, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.623631] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.623631] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1061.627705] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.917s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.717198] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Releasing lock "refresh_cache-3fabef9e-13ff-41b6-a4eb-ee65b66c37fe" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.717555] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Instance network_info: |[{"id": "f8c67797-fbfc-41d3-b441-9c5274d8c188", "address": "fa:16:3e:0a:72:7d", "network": {"id": "2ec6398f-b75e-44f0-a1c8-dffea7f87b26", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-177843661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78bce9f020ee429e9902f3fad3b62a3c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8c67797-fb", "ovs_interfaceid": "f8c67797-fbfc-41d3-b441-9c5274d8c188", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1061.718091] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0a:72:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cc0a33d-17c0-4b87-b48f-413a87a4cc6a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f8c67797-fbfc-41d3-b441-9c5274d8c188', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1061.725454] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Creating folder: Project (78bce9f020ee429e9902f3fad3b62a3c). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1061.729438] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-daecef74-5795-42cf-a6fe-66ee4e81a23d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.736643] env[61243]: DEBUG oslo_vmware.api [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339444, 'name': ReconfigVM_Task, 'duration_secs': 0.291822} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.736933] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Reconfigured VM instance instance-00000066 to attach disk [datastore2] bb7c810f-91d1-43eb-b0cc-fb4c686b1f29/bb7c810f-91d1-43eb-b0cc-fb4c686b1f29.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1061.737219] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updating instance 'bb7c810f-91d1-43eb-b0cc-fb4c686b1f29' progress to 50 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1061.741434] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Created folder: Project (78bce9f020ee429e9902f3fad3b62a3c) in parent group-v285636. [ 1061.741621] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Creating folder: Instances. Parent ref: group-v285786. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1061.741862] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b7cd22ae-4d22-4b46-98a1-e5bf492b769e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.751218] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Created folder: Instances in parent group-v285786. [ 1061.751218] env[61243]: DEBUG oslo.service.loopingcall [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1061.751218] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1061.751218] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3bb31949-ca12-4cf7-a939-3abd86a13279 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.768981] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1061.768981] env[61243]: value = "task-1339449" [ 1061.768981] env[61243]: _type = "Task" [ 1061.768981] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.778019] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339449, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.828394] env[61243]: DEBUG nova.compute.manager [req-7c65b2eb-2084-4542-b6b9-0e91afe473da req-d9463d24-67b7-4501-84ff-994f6cbb5c53 service nova] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Received event network-vif-deleted-b2f45b07-0ee1-45ba-a96b-9e8d62f5159d {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1061.828615] env[61243]: DEBUG nova.compute.manager [req-7c65b2eb-2084-4542-b6b9-0e91afe473da req-d9463d24-67b7-4501-84ff-994f6cbb5c53 service nova] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Received event network-vif-plugged-78eac2ea-b988-4138-a10a-a0fd00a3cdb1 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1061.828814] env[61243]: DEBUG oslo_concurrency.lockutils [req-7c65b2eb-2084-4542-b6b9-0e91afe473da req-d9463d24-67b7-4501-84ff-994f6cbb5c53 service nova] Acquiring lock "26ca8850-4785-4c02-9dd6-3788d0164510-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.829558] env[61243]: DEBUG oslo_concurrency.lockutils [req-7c65b2eb-2084-4542-b6b9-0e91afe473da req-d9463d24-67b7-4501-84ff-994f6cbb5c53 service nova] Lock "26ca8850-4785-4c02-9dd6-3788d0164510-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.829843] env[61243]: DEBUG oslo_concurrency.lockutils [req-7c65b2eb-2084-4542-b6b9-0e91afe473da req-d9463d24-67b7-4501-84ff-994f6cbb5c53 service nova] Lock "26ca8850-4785-4c02-9dd6-3788d0164510-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.830045] env[61243]: DEBUG nova.compute.manager [req-7c65b2eb-2084-4542-b6b9-0e91afe473da req-d9463d24-67b7-4501-84ff-994f6cbb5c53 service nova] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] No waiting events found dispatching network-vif-plugged-78eac2ea-b988-4138-a10a-a0fd00a3cdb1 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1061.830229] env[61243]: WARNING nova.compute.manager [req-7c65b2eb-2084-4542-b6b9-0e91afe473da req-d9463d24-67b7-4501-84ff-994f6cbb5c53 service nova] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Received unexpected event network-vif-plugged-78eac2ea-b988-4138-a10a-a0fd00a3cdb1 for instance with vm_state building and task_state spawning. [ 1061.844298] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Successfully updated port: 78eac2ea-b988-4138-a10a-a0fd00a3cdb1 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1061.894337] env[61243]: INFO nova.compute.manager [-] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Took 1.61 seconds to deallocate network for instance. [ 1061.988410] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339445, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072969} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.989822] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1061.993901] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6764b58f-bd52-4e5a-8195-f6321eafeea7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.998065] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.998370] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.004805] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339446, 'name': Rename_Task, 'duration_secs': 0.14128} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.014150] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1062.022872] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 5f1c246d-2732-4da4-a00b-b92e2990aa67/5f1c246d-2732-4da4-a00b-b92e2990aa67.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1062.023676] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-72ed77c6-64f6-46cd-a735-253c6bae9553 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.025470] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8e29d23-1785-4644-a556-4f1eb3d57ecb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.045860] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1062.045860] env[61243]: value = "task-1339451" [ 1062.045860] env[61243]: _type = "Task" [ 1062.045860] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.046159] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1062.046159] env[61243]: value = "task-1339450" [ 1062.046159] env[61243]: _type = "Task" [ 1062.046159] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.061316] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339450, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.061867] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339451, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.131476] env[61243]: DEBUG nova.compute.utils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1062.133611] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1062.133818] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1062.188130] env[61243]: DEBUG nova.policy [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ceff73e011724c0eaaa89b1012cabdfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78bce9f020ee429e9902f3fad3b62a3c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1062.247657] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7b8bd4-d868-4f24-a195-e3f104c4a6ee {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.273595] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a99edd6-76a1-4edd-8b9e-3effb304d1c7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.281567] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339449, 'name': CreateVM_Task, 'duration_secs': 0.360141} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.296500] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1062.296985] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updating instance 'bb7c810f-91d1-43eb-b0cc-fb4c686b1f29' progress to 67 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1062.300781] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.300912] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.301292] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1062.301750] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94df95e6-0868-472c-bcfd-456e171149df {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.306459] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1062.306459] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528f4da5-395f-d624-6a03-2bdf78679992" [ 1062.306459] env[61243]: _type = "Task" [ 1062.306459] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.314600] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528f4da5-395f-d624-6a03-2bdf78679992, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.346162] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "refresh_cache-26ca8850-4785-4c02-9dd6-3788d0164510" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.346356] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquired lock "refresh_cache-26ca8850-4785-4c02-9dd6-3788d0164510" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.346514] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1062.400647] env[61243]: DEBUG oslo_concurrency.lockutils [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.442789] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Successfully created port: b7573092-fe26-4633-ae34-f0b74fa0d3c7 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1062.502754] env[61243]: DEBUG nova.compute.utils [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1062.554670] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Volume attach. Driver type: vmdk {{(pid=61243) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1062.554911] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285783', 'volume_id': 'bce221b9-8c01-45eb-b91e-16b64854ec9c', 'name': 'volume-bce221b9-8c01-45eb-b91e-16b64854ec9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dca3301d-ef3e-48ef-920c-866b4a086ea6', 'attached_at': '', 'detached_at': '', 'volume_id': 'bce221b9-8c01-45eb-b91e-16b64854ec9c', 'serial': 'bce221b9-8c01-45eb-b91e-16b64854ec9c'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1062.556010] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e30b1a-671c-4f98-8912-bb9bd038ad6d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.563748] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339451, 'name': ReconfigVM_Task, 'duration_secs': 0.277421} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.576737] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 5f1c246d-2732-4da4-a00b-b92e2990aa67/5f1c246d-2732-4da4-a00b-b92e2990aa67.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1062.577894] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_type': 'disk', 'guest_format': None, 'boot_index': 0, 'encryption_format': None, 'encryption_secret_uuid': None, 'encrypted': False, 'disk_bus': None, 'encryption_options': None, 'size': 0, 'device_name': '/dev/sda', 'image_id': '6142e969-c114-4502-aa93-c018fb915a86'}], 'ephemerals': [], 'block_device_mapping': [{'guest_format': None, 'device_type': None, 'boot_index': None, 'mount_device': '/dev/sdb', 'attachment_id': 'b485ddf7-2023-49df-a7ea-88414c864e08', 'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285780', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'name': 'volume-d494ae0c-78b7-4745-85ac-83f826afe53a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5f1c246d-2732-4da4-a00b-b92e2990aa67', 'attached_at': '', 'detached_at': '', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'serial': 'd494ae0c-78b7-4745-85ac-83f826afe53a'}, 'delete_on_termination': False, 'volume_type': None}], 'swap': None} {{(pid=61243) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1062.578131] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Volume attach. Driver type: vmdk {{(pid=61243) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1062.578331] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285780', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'name': 'volume-d494ae0c-78b7-4745-85ac-83f826afe53a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5f1c246d-2732-4da4-a00b-b92e2990aa67', 'attached_at': '', 'detached_at': '', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'serial': 'd494ae0c-78b7-4745-85ac-83f826afe53a'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1062.578764] env[61243]: DEBUG oslo_vmware.api [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339450, 'name': PowerOnVM_Task, 'duration_secs': 0.500781} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.579488] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28fd7ba7-9b38-4a9c-84c8-74ae280b0010 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.582408] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e7f841-42e7-449c-8bb7-708c3704bd3a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.584698] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1062.585010] env[61243]: INFO nova.compute.manager [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Took 8.28 seconds to spawn the instance on the hypervisor. [ 1062.585112] env[61243]: DEBUG nova.compute.manager [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1062.585817] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f03df508-e6f0-4970-9cb5-0ae58cd01533 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.619558] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] volume-bce221b9-8c01-45eb-b91e-16b64854ec9c/volume-bce221b9-8c01-45eb-b91e-16b64854ec9c.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1062.621622] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a48a067d-ec6f-49b8-b796-e22f7bb74f6f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.636146] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b9a3324-cd29-4868-97c2-5d986245ce18 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.643571] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1062.646265] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Applying migration context for instance bb7c810f-91d1-43eb-b0cc-fb4c686b1f29 as it has an incoming, in-progress migration 0fcac3af-14cd-43cd-99d0-3e6f3647002e. Migration status is migrating {{(pid=61243) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1062.647784] env[61243]: INFO nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updating resource usage from migration 0fcac3af-14cd-43cd-99d0-3e6f3647002e [ 1062.667745] env[61243]: DEBUG oslo_vmware.api [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1062.667745] env[61243]: value = "task-1339452" [ 1062.667745] env[61243]: _type = "Task" [ 1062.667745] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.675363] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] volume-d494ae0c-78b7-4745-85ac-83f826afe53a/volume-d494ae0c-78b7-4745-85ac-83f826afe53a.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1062.676605] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 3f839aa5-2b9a-4807-b63b-931f74455532 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.676754] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 55f24903-b0b4-4124-afde-f31b2680a53a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.676883] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 5f1c246d-2732-4da4-a00b-b92e2990aa67 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.677013] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance dca3301d-ef3e-48ef-920c-866b4a086ea6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.677146] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 60fa2863-f0aa-4798-8e5f-e93df47f93e8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.677266] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 466f6a0a-9bc2-4709-8625-e6fc65b68f80 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.677403] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 486ebf65-4376-424c-9deb-f1ef70e44be3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.677520] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 8279496d-4a90-4962-9cc2-5d5e1ee33ff0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.677635] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance df141b12-1d34-4a47-b6bb-15cb75d266c3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.677750] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Migration 0fcac3af-14cd-43cd-99d0-3e6f3647002e is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1062.677864] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance bb7c810f-91d1-43eb-b0cc-fb4c686b1f29 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.677976] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.678110] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 26ca8850-4785-4c02-9dd6-3788d0164510 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.678227] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance fbf413f0-389c-4924-bf5d-400f95c89a84 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.678436] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1062.678572] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3264MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1062.681114] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-19594059-077c-44ec-85b0-8fe8d6952524 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.707661] env[61243]: DEBUG oslo_vmware.api [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339452, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.708673] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1062.708673] env[61243]: value = "task-1339453" [ 1062.708673] env[61243]: _type = "Task" [ 1062.708673] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.717954] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339453, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.817669] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]528f4da5-395f-d624-6a03-2bdf78679992, 'name': SearchDatastore_Task, 'duration_secs': 0.012907} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.820782] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.821062] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1062.821342] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.821500] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.821715] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1062.822256] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b816bfaa-0e88-4d7f-b99c-73c17de69cc7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.830770] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1062.830957] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1062.831744] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08eec367-91ff-416b-9ccf-b2441b91555b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.838057] env[61243]: DEBUG nova.network.neutron [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Port a968f6e7-b39a-4dd8-abf8-bf0919565a72 binding to destination host cpu-1 is already ACTIVE {{(pid=61243) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1062.841328] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1062.841328] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a189cc-63c9-c058-6073-bf0cacfc6703" [ 1062.841328] env[61243]: _type = "Task" [ 1062.841328] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.853822] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a189cc-63c9-c058-6073-bf0cacfc6703, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.885631] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1062.917563] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6213c86-f36e-4864-800c-94839dc1d4ae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.922230] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6f3384-4754-4998-a74f-cac9cf48c4b8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.955627] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f359b492-8d35-4fe8-8180-3b4b647d311d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.961988] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a721eb-6a1c-4725-ad07-0f4493b96797 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.976871] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1063.007952] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.053513] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Updating instance_info_cache with network_info: [{"id": "78eac2ea-b988-4138-a10a-a0fd00a3cdb1", "address": "fa:16:3e:67:21:db", "network": {"id": "2ec6398f-b75e-44f0-a1c8-dffea7f87b26", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-177843661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78bce9f020ee429e9902f3fad3b62a3c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78eac2ea-b9", "ovs_interfaceid": "78eac2ea-b988-4138-a10a-a0fd00a3cdb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.149344] env[61243]: INFO nova.compute.manager [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Took 13.31 seconds to build instance. [ 1063.190534] env[61243]: DEBUG oslo_vmware.api [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339452, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.217488] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339453, 'name': ReconfigVM_Task, 'duration_secs': 0.327023} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.217713] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Reconfigured VM instance instance-0000005e to attach disk [datastore1] volume-d494ae0c-78b7-4745-85ac-83f826afe53a/volume-d494ae0c-78b7-4745-85ac-83f826afe53a.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1063.222302] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14834ebf-f850-4a9a-b4ac-622e57f573de {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.236624] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1063.236624] env[61243]: value = "task-1339454" [ 1063.236624] env[61243]: _type = "Task" [ 1063.236624] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.244660] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339454, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.353756] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a189cc-63c9-c058-6073-bf0cacfc6703, 'name': SearchDatastore_Task, 'duration_secs': 0.031185} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.354563] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed12b2a9-47c0-418e-b01c-3e9495f464cf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.360665] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1063.360665] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e13515-62a6-3450-9abd-d4ee4f96072f" [ 1063.360665] env[61243]: _type = "Task" [ 1063.360665] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.368569] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e13515-62a6-3450-9abd-d4ee4f96072f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.500936] env[61243]: ERROR nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [req-2ed28a69-c31a-449c-b1b2-4dc530f3561c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2ed28a69-c31a-449c-b1b2-4dc530f3561c"}]} [ 1063.516961] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1063.532585] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1063.532775] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1063.543429] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1063.554867] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Releasing lock "refresh_cache-26ca8850-4785-4c02-9dd6-3788d0164510" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.555852] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Instance network_info: |[{"id": "78eac2ea-b988-4138-a10a-a0fd00a3cdb1", "address": "fa:16:3e:67:21:db", "network": {"id": "2ec6398f-b75e-44f0-a1c8-dffea7f87b26", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-177843661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78bce9f020ee429e9902f3fad3b62a3c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78eac2ea-b9", "ovs_interfaceid": "78eac2ea-b988-4138-a10a-a0fd00a3cdb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1063.555852] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:21:db', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cc0a33d-17c0-4b87-b48f-413a87a4cc6a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '78eac2ea-b988-4138-a10a-a0fd00a3cdb1', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1063.563284] env[61243]: DEBUG oslo.service.loopingcall [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1063.564233] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1063.567165] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1063.567617] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-477d2934-5140-4e22-a013-0744526f1fe3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.588530] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1063.588530] env[61243]: value = "task-1339455" [ 1063.588530] env[61243]: _type = "Task" [ 1063.588530] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.596407] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339455, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.651080] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f8a25a8a-ab18-42a7-91c2-06dd60a12cfb tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "df141b12-1d34-4a47-b6bb-15cb75d266c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.827s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.672327] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1063.692358] env[61243]: DEBUG oslo_vmware.api [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339452, 'name': ReconfigVM_Task, 'duration_secs': 0.715418} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.696494] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Reconfigured VM instance instance-00000060 to attach disk [datastore1] volume-bce221b9-8c01-45eb-b91e-16b64854ec9c/volume-bce221b9-8c01-45eb-b91e-16b64854ec9c.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1063.701913] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62a7418e-d214-4cf9-93fa-3cb243ba2758 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.714465] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1063.714708] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1063.714886] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1063.715125] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1063.715290] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1063.715455] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1063.715674] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1063.715831] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1063.716009] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1063.716185] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1063.716366] env[61243]: DEBUG nova.virt.hardware [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1063.719958] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a48e4ec7-4319-4c44-89f7-0e8dba7bff3d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.728719] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aa77707-685d-4389-b95d-a32d0691acc8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.735622] env[61243]: DEBUG oslo_vmware.api [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1063.735622] env[61243]: value = "task-1339456" [ 1063.735622] env[61243]: _type = "Task" [ 1063.735622] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.759061] env[61243]: DEBUG oslo_vmware.api [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339456, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.761699] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339454, 'name': ReconfigVM_Task, 'duration_secs': 0.170329} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.762016] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285780', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'name': 'volume-d494ae0c-78b7-4745-85ac-83f826afe53a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5f1c246d-2732-4da4-a00b-b92e2990aa67', 'attached_at': '', 'detached_at': '', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'serial': 'd494ae0c-78b7-4745-85ac-83f826afe53a'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1063.762631] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ff934c5b-c313-4338-b7ed-122ef9a6eff3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.765630] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f8c662-e998-44f5-ac28-4279a671bbb1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.774972] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1086d921-ddaf-48e0-9fe8-a2776260667e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.777220] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1063.777220] env[61243]: value = "task-1339457" [ 1063.777220] env[61243]: _type = "Task" [ 1063.777220] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.806233] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa263d6-d1b8-44b4-9c04-fcf4a986c403 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.811625] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339457, 'name': Rename_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.817139] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a999ae4c-a88a-44e3-8aaa-afae34395f10 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.830814] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1063.860684] env[61243]: DEBUG nova.compute.manager [req-4a9c9807-fb66-4afc-b9e4-63d4aedb72ed req-3daca9a3-bb46-4692-9817-d6f5f12bd5cb service nova] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Received event network-changed-78eac2ea-b988-4138-a10a-a0fd00a3cdb1 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1063.860903] env[61243]: DEBUG nova.compute.manager [req-4a9c9807-fb66-4afc-b9e4-63d4aedb72ed req-3daca9a3-bb46-4692-9817-d6f5f12bd5cb service nova] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Refreshing instance network info cache due to event network-changed-78eac2ea-b988-4138-a10a-a0fd00a3cdb1. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1063.861191] env[61243]: DEBUG oslo_concurrency.lockutils [req-4a9c9807-fb66-4afc-b9e4-63d4aedb72ed req-3daca9a3-bb46-4692-9817-d6f5f12bd5cb service nova] Acquiring lock "refresh_cache-26ca8850-4785-4c02-9dd6-3788d0164510" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.861374] env[61243]: DEBUG oslo_concurrency.lockutils [req-4a9c9807-fb66-4afc-b9e4-63d4aedb72ed req-3daca9a3-bb46-4692-9817-d6f5f12bd5cb service nova] Acquired lock "refresh_cache-26ca8850-4785-4c02-9dd6-3788d0164510" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.861548] env[61243]: DEBUG nova.network.neutron [req-4a9c9807-fb66-4afc-b9e4-63d4aedb72ed req-3daca9a3-bb46-4692-9817-d6f5f12bd5cb service nova] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Refreshing network info cache for port 78eac2ea-b988-4138-a10a-a0fd00a3cdb1 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1063.869054] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.869054] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.869236] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.881065] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e13515-62a6-3450-9abd-d4ee4f96072f, 'name': SearchDatastore_Task, 'duration_secs': 0.01093} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.881368] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.881943] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe/3fabef9e-13ff-41b6-a4eb-ee65b66c37fe.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1063.882851] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e779bcfe-0779-4d4f-86aa-ad7772e01312 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.889469] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1063.889469] env[61243]: value = "task-1339458" [ 1063.889469] env[61243]: _type = "Task" [ 1063.889469] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.898791] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339458, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.967682] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Successfully updated port: b7573092-fe26-4633-ae34-f0b74fa0d3c7 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1064.053448] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.053710] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.053950] env[61243]: INFO nova.compute.manager [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Attaching volume 77ec08cb-5d18-4bee-b234-4aca1c5ce999 to /dev/sdb [ 1064.086698] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3930e792-824f-4927-b527-bf74897fd474 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.099788] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8758ec-6813-43d6-b89f-659806b539e0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.102175] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339455, 'name': CreateVM_Task, 'duration_secs': 0.497024} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.103194] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1064.103442] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.103610] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.104134] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1064.104204] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6b6b0a4-4b90-4257-990e-47fc8a68aa17 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.108499] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1064.108499] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529be3e1-cfab-0035-c849-860ff4ce9c82" [ 1064.108499] env[61243]: _type = "Task" [ 1064.108499] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.115729] env[61243]: DEBUG nova.virt.block_device [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating existing volume attachment record: 21a8bae9-1ca3-44d5-a2a2-d145826d3a18 {{(pid=61243) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1064.124100] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529be3e1-cfab-0035-c849-860ff4ce9c82, 'name': SearchDatastore_Task, 'duration_secs': 0.011235} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.124393] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.124672] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1064.124855] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.125011] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.125259] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1064.125556] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a7bc82ea-1cff-4501-aadd-22ae9cbdd0fb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.133897] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1064.134030] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1064.134790] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f70756d-de49-409c-a4f3-717393f4f35c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.142834] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1064.142834] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a0710b-88b1-21f6-cb00-19e5f27116de" [ 1064.142834] env[61243]: _type = "Task" [ 1064.142834] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.154025] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a0710b-88b1-21f6-cb00-19e5f27116de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.246215] env[61243]: DEBUG oslo_vmware.api [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339456, 'name': ReconfigVM_Task, 'duration_secs': 0.173082} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.246633] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285783', 'volume_id': 'bce221b9-8c01-45eb-b91e-16b64854ec9c', 'name': 'volume-bce221b9-8c01-45eb-b91e-16b64854ec9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dca3301d-ef3e-48ef-920c-866b4a086ea6', 'attached_at': '', 'detached_at': '', 'volume_id': 'bce221b9-8c01-45eb-b91e-16b64854ec9c', 'serial': 'bce221b9-8c01-45eb-b91e-16b64854ec9c'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1064.287584] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339457, 'name': Rename_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.364830] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 153 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1064.366118] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 153 to 154 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1064.366118] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1064.403020] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339458, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.469992] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "refresh_cache-fbf413f0-389c-4924-bf5d-400f95c89a84" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.470310] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquired lock "refresh_cache-fbf413f0-389c-4924-bf5d-400f95c89a84" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.470502] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1064.590104] env[61243]: DEBUG nova.network.neutron [req-4a9c9807-fb66-4afc-b9e4-63d4aedb72ed req-3daca9a3-bb46-4692-9817-d6f5f12bd5cb service nova] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Updated VIF entry in instance network info cache for port 78eac2ea-b988-4138-a10a-a0fd00a3cdb1. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1064.590420] env[61243]: DEBUG nova.network.neutron [req-4a9c9807-fb66-4afc-b9e4-63d4aedb72ed req-3daca9a3-bb46-4692-9817-d6f5f12bd5cb service nova] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Updating instance_info_cache with network_info: [{"id": "78eac2ea-b988-4138-a10a-a0fd00a3cdb1", "address": "fa:16:3e:67:21:db", "network": {"id": "2ec6398f-b75e-44f0-a1c8-dffea7f87b26", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-177843661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78bce9f020ee429e9902f3fad3b62a3c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78eac2ea-b9", "ovs_interfaceid": "78eac2ea-b988-4138-a10a-a0fd00a3cdb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.653786] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a0710b-88b1-21f6-cb00-19e5f27116de, 'name': SearchDatastore_Task, 'duration_secs': 0.079637} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.654433] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a0bba6f-1ed0-4944-a2e9-d71d093fbad8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.659260] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1064.659260] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e569c3-d6f5-e6de-46b0-6724f7543f0a" [ 1064.659260] env[61243]: _type = "Task" [ 1064.659260] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.666738] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e569c3-d6f5-e6de-46b0-6724f7543f0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.789531] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339457, 'name': Rename_Task, 'duration_secs': 0.845493} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.789531] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1064.789531] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d0657ffe-d8e4-42d4-8b41-b983886ac4f2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.796174] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1064.796174] env[61243]: value = "task-1339462" [ 1064.796174] env[61243]: _type = "Task" [ 1064.796174] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.803746] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339462, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.810519] env[61243]: DEBUG oslo_concurrency.lockutils [None req-524996f0-2e5d-4c39-8594-dabf9885a261 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "df141b12-1d34-4a47-b6bb-15cb75d266c3" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.810753] env[61243]: DEBUG oslo_concurrency.lockutils [None req-524996f0-2e5d-4c39-8594-dabf9885a261 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "df141b12-1d34-4a47-b6bb-15cb75d266c3" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.810935] env[61243]: DEBUG nova.compute.manager [None req-524996f0-2e5d-4c39-8594-dabf9885a261 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1064.811703] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0826f34f-ab3a-4e14-a726-27a6cab15de1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.817496] env[61243]: DEBUG nova.compute.manager [None req-524996f0-2e5d-4c39-8594-dabf9885a261 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61243) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1064.818065] env[61243]: DEBUG nova.objects.instance [None req-524996f0-2e5d-4c39-8594-dabf9885a261 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lazy-loading 'flavor' on Instance uuid df141b12-1d34-4a47-b6bb-15cb75d266c3 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.871670] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61243) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1064.872056] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.244s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.872361] env[61243]: DEBUG oslo_concurrency.lockutils [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.472s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.872725] env[61243]: DEBUG nova.objects.instance [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lazy-loading 'resources' on Instance uuid 55f24903-b0b4-4124-afde-f31b2680a53a {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.899989] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339458, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.651975} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.900290] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe/3fabef9e-13ff-41b6-a4eb-ee65b66c37fe.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1064.900511] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1064.900765] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f9249eb8-74f4-482d-921c-77441d77aeac {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.904121] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "refresh_cache-bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.904309] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "refresh_cache-bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.904483] env[61243]: DEBUG nova.network.neutron [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1064.907435] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1064.907435] env[61243]: value = "task-1339463" [ 1064.907435] env[61243]: _type = "Task" [ 1064.907435] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.915214] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339463, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.008218] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1065.094031] env[61243]: DEBUG oslo_concurrency.lockutils [req-4a9c9807-fb66-4afc-b9e4-63d4aedb72ed req-3daca9a3-bb46-4692-9817-d6f5f12bd5cb service nova] Releasing lock "refresh_cache-26ca8850-4785-4c02-9dd6-3788d0164510" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.153348] env[61243]: DEBUG nova.network.neutron [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Updating instance_info_cache with network_info: [{"id": "b7573092-fe26-4633-ae34-f0b74fa0d3c7", "address": "fa:16:3e:57:6c:ae", "network": {"id": "2ec6398f-b75e-44f0-a1c8-dffea7f87b26", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-177843661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78bce9f020ee429e9902f3fad3b62a3c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7573092-fe", "ovs_interfaceid": "b7573092-fe26-4633-ae34-f0b74fa0d3c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.169756] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e569c3-d6f5-e6de-46b0-6724f7543f0a, 'name': SearchDatastore_Task, 'duration_secs': 0.008418} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.170122] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.170305] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 26ca8850-4785-4c02-9dd6-3788d0164510/26ca8850-4785-4c02-9dd6-3788d0164510.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1065.170588] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-94061219-a148-4986-82bb-ff102acd1e13 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.177325] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1065.177325] env[61243]: value = "task-1339464" [ 1065.177325] env[61243]: _type = "Task" [ 1065.177325] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.186145] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339464, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.285476] env[61243]: DEBUG nova.objects.instance [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lazy-loading 'flavor' on Instance uuid dca3301d-ef3e-48ef-920c-866b4a086ea6 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.308811] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339462, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.323490] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-524996f0-2e5d-4c39-8594-dabf9885a261 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1065.323821] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e9003d78-1459-46b9-ac0b-9d95b1be2244 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.330109] env[61243]: DEBUG oslo_vmware.api [None req-524996f0-2e5d-4c39-8594-dabf9885a261 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1065.330109] env[61243]: value = "task-1339465" [ 1065.330109] env[61243]: _type = "Task" [ 1065.330109] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.338687] env[61243]: DEBUG oslo_vmware.api [None req-524996f0-2e5d-4c39-8594-dabf9885a261 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339465, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.421279] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339463, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066952} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.421621] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1065.422587] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e051280-b93e-4b0e-a105-8944f8c9e4e0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.452879] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe/3fabef9e-13ff-41b6-a4eb-ee65b66c37fe.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1065.457911] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f71b2bb-e3ac-46ca-b2d4-59b997b43505 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.478518] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1065.478518] env[61243]: value = "task-1339466" [ 1065.478518] env[61243]: _type = "Task" [ 1065.478518] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.489807] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339466, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.632602] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79ea46df-2cf5-410f-88a6-5f09bc7137a1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.640746] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c85f910a-f724-4ef2-886e-191c09b3e243 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.671842] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Releasing lock "refresh_cache-fbf413f0-389c-4924-bf5d-400f95c89a84" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.672200] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Instance network_info: |[{"id": "b7573092-fe26-4633-ae34-f0b74fa0d3c7", "address": "fa:16:3e:57:6c:ae", "network": {"id": "2ec6398f-b75e-44f0-a1c8-dffea7f87b26", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-177843661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78bce9f020ee429e9902f3fad3b62a3c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7573092-fe", "ovs_interfaceid": "b7573092-fe26-4633-ae34-f0b74fa0d3c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1065.673656] env[61243]: DEBUG nova.network.neutron [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updating instance_info_cache with network_info: [{"id": "a968f6e7-b39a-4dd8-abf8-bf0919565a72", "address": "fa:16:3e:80:14:77", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa968f6e7-b3", "ovs_interfaceid": "a968f6e7-b39a-4dd8-abf8-bf0919565a72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.675058] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:57:6c:ae', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cc0a33d-17c0-4b87-b48f-413a87a4cc6a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b7573092-fe26-4633-ae34-f0b74fa0d3c7', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1065.683506] env[61243]: DEBUG oslo.service.loopingcall [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1065.683792] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4741d00e-2e33-4bb1-9d14-310eea8a6796 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.687667] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "refresh_cache-bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.691299] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1065.695298] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d1d9f74f-0f28-4c7b-a8d0-0a222fec8002 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.721485] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339464, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.724679] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1065.724679] env[61243]: value = "task-1339467" [ 1065.724679] env[61243]: _type = "Task" [ 1065.724679] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.725950] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3224284c-bc23-4f78-baee-eed016088dd6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.745934] env[61243]: DEBUG nova.compute.provider_tree [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.755814] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339467, 'name': CreateVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.792035] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0a8b6b16-a07d-4b85-bf41-c27ff11085c5 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "dca3301d-ef3e-48ef-920c-866b4a086ea6" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 9.872s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.806221] env[61243]: DEBUG oslo_vmware.api [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339462, 'name': PowerOnVM_Task, 'duration_secs': 0.687449} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.807021] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1065.807273] env[61243]: DEBUG nova.compute.manager [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.808030] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c6b2869-3848-456f-9b3f-e084520c7015 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.839100] env[61243]: DEBUG oslo_vmware.api [None req-524996f0-2e5d-4c39-8594-dabf9885a261 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339465, 'name': PowerOffVM_Task, 'duration_secs': 0.291512} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.839382] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-524996f0-2e5d-4c39-8594-dabf9885a261 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1065.839552] env[61243]: DEBUG nova.compute.manager [None req-524996f0-2e5d-4c39-8594-dabf9885a261 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.840332] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524b550e-523a-453d-b183-099819467649 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.887681] env[61243]: DEBUG nova.compute.manager [req-04c1c1f9-b974-497b-b9e0-6694796a00c5 req-d6ef17b0-1e66-4b66-bc30-404b81844cb8 service nova] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Received event network-vif-plugged-b7573092-fe26-4633-ae34-f0b74fa0d3c7 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1065.887746] env[61243]: DEBUG oslo_concurrency.lockutils [req-04c1c1f9-b974-497b-b9e0-6694796a00c5 req-d6ef17b0-1e66-4b66-bc30-404b81844cb8 service nova] Acquiring lock "fbf413f0-389c-4924-bf5d-400f95c89a84-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.887977] env[61243]: DEBUG oslo_concurrency.lockutils [req-04c1c1f9-b974-497b-b9e0-6694796a00c5 req-d6ef17b0-1e66-4b66-bc30-404b81844cb8 service nova] Lock "fbf413f0-389c-4924-bf5d-400f95c89a84-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.888229] env[61243]: DEBUG oslo_concurrency.lockutils [req-04c1c1f9-b974-497b-b9e0-6694796a00c5 req-d6ef17b0-1e66-4b66-bc30-404b81844cb8 service nova] Lock "fbf413f0-389c-4924-bf5d-400f95c89a84-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.888435] env[61243]: DEBUG nova.compute.manager [req-04c1c1f9-b974-497b-b9e0-6694796a00c5 req-d6ef17b0-1e66-4b66-bc30-404b81844cb8 service nova] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] No waiting events found dispatching network-vif-plugged-b7573092-fe26-4633-ae34-f0b74fa0d3c7 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1065.888656] env[61243]: WARNING nova.compute.manager [req-04c1c1f9-b974-497b-b9e0-6694796a00c5 req-d6ef17b0-1e66-4b66-bc30-404b81844cb8 service nova] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Received unexpected event network-vif-plugged-b7573092-fe26-4633-ae34-f0b74fa0d3c7 for instance with vm_state building and task_state spawning. [ 1065.888830] env[61243]: DEBUG nova.compute.manager [req-04c1c1f9-b974-497b-b9e0-6694796a00c5 req-d6ef17b0-1e66-4b66-bc30-404b81844cb8 service nova] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Received event network-changed-b7573092-fe26-4633-ae34-f0b74fa0d3c7 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1065.889031] env[61243]: DEBUG nova.compute.manager [req-04c1c1f9-b974-497b-b9e0-6694796a00c5 req-d6ef17b0-1e66-4b66-bc30-404b81844cb8 service nova] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Refreshing instance network info cache due to event network-changed-b7573092-fe26-4633-ae34-f0b74fa0d3c7. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1065.889288] env[61243]: DEBUG oslo_concurrency.lockutils [req-04c1c1f9-b974-497b-b9e0-6694796a00c5 req-d6ef17b0-1e66-4b66-bc30-404b81844cb8 service nova] Acquiring lock "refresh_cache-fbf413f0-389c-4924-bf5d-400f95c89a84" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.889453] env[61243]: DEBUG oslo_concurrency.lockutils [req-04c1c1f9-b974-497b-b9e0-6694796a00c5 req-d6ef17b0-1e66-4b66-bc30-404b81844cb8 service nova] Acquired lock "refresh_cache-fbf413f0-389c-4924-bf5d-400f95c89a84" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.889620] env[61243]: DEBUG nova.network.neutron [req-04c1c1f9-b974-497b-b9e0-6694796a00c5 req-d6ef17b0-1e66-4b66-bc30-404b81844cb8 service nova] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Refreshing network info cache for port b7573092-fe26-4633-ae34-f0b74fa0d3c7 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1065.987958] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339466, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.205475] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339464, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.895797} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.205475] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 26ca8850-4785-4c02-9dd6-3788d0164510/26ca8850-4785-4c02-9dd6-3788d0164510.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1066.205475] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1066.205940] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e9ec005e-2696-46aa-a83a-0047e624668e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.212546] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1066.212546] env[61243]: value = "task-1339468" [ 1066.212546] env[61243]: _type = "Task" [ 1066.212546] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.218954] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75681898-d77a-460b-ab2f-7117b7861ba7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.226444] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339468, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.243856] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a043dd7c-66d5-4967-aaf3-2b2bbd6a91fa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.252175] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339467, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.254798] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updating instance 'bb7c810f-91d1-43eb-b0cc-fb4c686b1f29' progress to 83 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1066.259077] env[61243]: DEBUG nova.scheduler.client.report [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1066.262185] env[61243]: DEBUG oslo_concurrency.lockutils [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "dca3301d-ef3e-48ef-920c-866b4a086ea6" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.263029] env[61243]: DEBUG oslo_concurrency.lockutils [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "dca3301d-ef3e-48ef-920c-866b4a086ea6" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.325983] env[61243]: DEBUG oslo_concurrency.lockutils [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.352945] env[61243]: DEBUG oslo_concurrency.lockutils [None req-524996f0-2e5d-4c39-8594-dabf9885a261 tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "df141b12-1d34-4a47-b6bb-15cb75d266c3" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.542s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.490391] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339466, 'name': ReconfigVM_Task, 'duration_secs': 0.626646} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.490391] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe/3fabef9e-13ff-41b6-a4eb-ee65b66c37fe.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1066.492392] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-191cbffb-64d7-4e86-97c9-fb64b652a247 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.498981] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1066.498981] env[61243]: value = "task-1339469" [ 1066.498981] env[61243]: _type = "Task" [ 1066.498981] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.508432] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339469, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.590265] env[61243]: DEBUG nova.network.neutron [req-04c1c1f9-b974-497b-b9e0-6694796a00c5 req-d6ef17b0-1e66-4b66-bc30-404b81844cb8 service nova] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Updated VIF entry in instance network info cache for port b7573092-fe26-4633-ae34-f0b74fa0d3c7. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1066.590688] env[61243]: DEBUG nova.network.neutron [req-04c1c1f9-b974-497b-b9e0-6694796a00c5 req-d6ef17b0-1e66-4b66-bc30-404b81844cb8 service nova] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Updating instance_info_cache with network_info: [{"id": "b7573092-fe26-4633-ae34-f0b74fa0d3c7", "address": "fa:16:3e:57:6c:ae", "network": {"id": "2ec6398f-b75e-44f0-a1c8-dffea7f87b26", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-177843661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78bce9f020ee429e9902f3fad3b62a3c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7573092-fe", "ovs_interfaceid": "b7573092-fe26-4633-ae34-f0b74fa0d3c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.722176] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339468, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070296} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.722448] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1066.723256] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c7b435-402b-478c-909b-fa3679093b3e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.744913] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 26ca8850-4785-4c02-9dd6-3788d0164510/26ca8850-4785-4c02-9dd6-3788d0164510.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1066.745450] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04a6437a-b2ee-4b8d-be27-eb1584ae3f57 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.765485] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1066.766715] env[61243]: DEBUG oslo_concurrency.lockutils [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.894s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.768313] env[61243]: INFO nova.compute.manager [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Detaching volume bce221b9-8c01-45eb-b91e-16b64854ec9c [ 1066.773898] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-931df234-dce2-448e-ab9d-6fd0308150e7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.775748] env[61243]: DEBUG oslo_concurrency.lockutils [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.450s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.775748] env[61243]: DEBUG nova.objects.instance [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61243) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1066.778952] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1066.778952] env[61243]: value = "task-1339471" [ 1066.778952] env[61243]: _type = "Task" [ 1066.778952] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.778952] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339467, 'name': CreateVM_Task, 'duration_secs': 0.661277} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.781280] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1066.786242] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.786424] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.786750] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1066.787895] env[61243]: INFO nova.scheduler.client.report [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Deleted allocations for instance 55f24903-b0b4-4124-afde-f31b2680a53a [ 1066.790031] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67f7b324-1003-4b6e-b934-fce4cdebae30 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.792084] env[61243]: DEBUG oslo_vmware.api [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1066.792084] env[61243]: value = "task-1339472" [ 1066.792084] env[61243]: _type = "Task" [ 1066.792084] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.805654] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339471, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.805654] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1066.805654] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52674779-e298-e286-723f-95e37face1ff" [ 1066.805654] env[61243]: _type = "Task" [ 1066.805654] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.811887] env[61243]: DEBUG oslo_vmware.api [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339472, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.818311] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52674779-e298-e286-723f-95e37face1ff, 'name': SearchDatastore_Task, 'duration_secs': 0.009022} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.818311] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.818311] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1066.818534] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.818534] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.819345] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1066.819791] env[61243]: INFO nova.virt.block_device [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Attempting to driver detach volume bce221b9-8c01-45eb-b91e-16b64854ec9c from mountpoint /dev/sdb [ 1066.820017] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Volume detach. Driver type: vmdk {{(pid=61243) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1066.820220] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285783', 'volume_id': 'bce221b9-8c01-45eb-b91e-16b64854ec9c', 'name': 'volume-bce221b9-8c01-45eb-b91e-16b64854ec9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dca3301d-ef3e-48ef-920c-866b4a086ea6', 'attached_at': '', 'detached_at': '', 'volume_id': 'bce221b9-8c01-45eb-b91e-16b64854ec9c', 'serial': 'bce221b9-8c01-45eb-b91e-16b64854ec9c'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1066.820496] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cf3e0987-2086-497a-8e68-c3551253e561 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.822923] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd90300e-37cd-4ec7-8665-362bb2032cf5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.845008] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b903443-184f-4061-9c76-105c0ba44935 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.847680] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1066.847868] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1066.849023] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bb48075-1d90-4a9f-924b-f10046ab5aaa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.854855] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1066.854855] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523bc34f-8a15-51d5-f7c9-c9c4b610577d" [ 1066.854855] env[61243]: _type = "Task" [ 1066.854855] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.859986] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d665c01-e167-42e5-bf14-088f4e3b50dd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.867659] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]523bc34f-8a15-51d5-f7c9-c9c4b610577d, 'name': SearchDatastore_Task, 'duration_secs': 0.008962} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.883578] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8a5a185-d8f2-43d1-8bb4-84eec6a83bf1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.886366] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e8dfb6-5a4d-4b5f-89c9-2dca004485e8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.891671] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1066.891671] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ab805a-1852-492e-e5f4-6408d1e9580e" [ 1066.891671] env[61243]: _type = "Task" [ 1066.891671] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.904110] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] The volume has not been displaced from its original location: [datastore1] volume-bce221b9-8c01-45eb-b91e-16b64854ec9c/volume-bce221b9-8c01-45eb-b91e-16b64854ec9c.vmdk. No consolidation needed. {{(pid=61243) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1066.909196] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Reconfiguring VM instance instance-00000060 to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1066.912142] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-089f552a-1eeb-4407-82cd-ec0d38e44214 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.929757] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ab805a-1852-492e-e5f4-6408d1e9580e, 'name': SearchDatastore_Task, 'duration_secs': 0.008569} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.930956] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.931191] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] fbf413f0-389c-4924-bf5d-400f95c89a84/fbf413f0-389c-4924-bf5d-400f95c89a84.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1066.931507] env[61243]: DEBUG oslo_vmware.api [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1066.931507] env[61243]: value = "task-1339473" [ 1066.931507] env[61243]: _type = "Task" [ 1066.931507] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.931695] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a6a6226a-a9dd-48c4-a64a-1eaf425da78b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.941605] env[61243]: DEBUG oslo_vmware.api [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339473, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.942743] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1066.942743] env[61243]: value = "task-1339474" [ 1066.942743] env[61243]: _type = "Task" [ 1066.942743] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.949692] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339474, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.009871] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339469, 'name': Rename_Task, 'duration_secs': 0.134638} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.010577] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1067.010577] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eca4324a-60c5-4bc5-9635-349e08e95486 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.018926] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1067.018926] env[61243]: value = "task-1339475" [ 1067.018926] env[61243]: _type = "Task" [ 1067.018926] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.028193] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339475, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.093528] env[61243]: DEBUG oslo_concurrency.lockutils [req-04c1c1f9-b974-497b-b9e0-6694796a00c5 req-d6ef17b0-1e66-4b66-bc30-404b81844cb8 service nova] Releasing lock "refresh_cache-fbf413f0-389c-4924-bf5d-400f95c89a84" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.200482] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "5f1c246d-2732-4da4-a00b-b92e2990aa67" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.201012] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "5f1c246d-2732-4da4-a00b-b92e2990aa67" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.300335] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339471, 'name': ReconfigVM_Task, 'duration_secs': 0.30601} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.300833] env[61243]: DEBUG oslo_concurrency.lockutils [None req-060313a6-64d5-4e8d-ada4-2ff296eb6bfb tempest-AttachVolumeTestJSON-1707899569 tempest-AttachVolumeTestJSON-1707899569-project-member] Lock "55f24903-b0b4-4124-afde-f31b2680a53a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.146s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.306117] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 26ca8850-4785-4c02-9dd6-3788d0164510/26ca8850-4785-4c02-9dd6-3788d0164510.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1067.307029] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9fe207d6-d8ca-48b3-bcd1-f5744dc0c6dd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.317497] env[61243]: DEBUG oslo_vmware.api [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339472, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.319755] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1067.319755] env[61243]: value = "task-1339476" [ 1067.319755] env[61243]: _type = "Task" [ 1067.319755] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.330207] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339476, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.444987] env[61243]: DEBUG oslo_vmware.api [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339473, 'name': ReconfigVM_Task, 'duration_secs': 0.228954} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.449101] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Reconfigured VM instance instance-00000060 to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1067.454763] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6849d85-a8f3-49ba-b3da-9d5519b83e4a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.472522] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339474, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.474204] env[61243]: DEBUG oslo_vmware.api [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1067.474204] env[61243]: value = "task-1339477" [ 1067.474204] env[61243]: _type = "Task" [ 1067.474204] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.483719] env[61243]: DEBUG oslo_vmware.api [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339477, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.533324] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339475, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.664487] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "df141b12-1d34-4a47-b6bb-15cb75d266c3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.664694] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "df141b12-1d34-4a47-b6bb-15cb75d266c3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.664941] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "df141b12-1d34-4a47-b6bb-15cb75d266c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.665197] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "df141b12-1d34-4a47-b6bb-15cb75d266c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.665396] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "df141b12-1d34-4a47-b6bb-15cb75d266c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.668383] env[61243]: INFO nova.compute.manager [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Terminating instance [ 1067.670719] env[61243]: DEBUG nova.compute.manager [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1067.671059] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1067.672290] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467b8cf1-4ee3-4ec9-b128-dd6481895dfb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.680841] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1067.681143] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-252dc94a-8783-45ae-98d7-d883edc61d7a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.704258] env[61243]: INFO nova.compute.manager [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Detaching volume d494ae0c-78b7-4745-85ac-83f826afe53a [ 1067.738606] env[61243]: INFO nova.virt.block_device [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Attempting to driver detach volume d494ae0c-78b7-4745-85ac-83f826afe53a from mountpoint /dev/sdb [ 1067.738888] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Volume detach. Driver type: vmdk {{(pid=61243) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1067.739135] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285780', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'name': 'volume-d494ae0c-78b7-4745-85ac-83f826afe53a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5f1c246d-2732-4da4-a00b-b92e2990aa67', 'attached_at': '', 'detached_at': '', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'serial': 'd494ae0c-78b7-4745-85ac-83f826afe53a'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1067.740134] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-251cd0a2-9604-4dc2-9c4c-75cfbd01380b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.762618] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f2c8be-6a07-4d59-aed5-c3a00fe8c204 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.767593] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1067.767888] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1067.768143] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleting the datastore file [datastore1] df141b12-1d34-4a47-b6bb-15cb75d266c3 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1067.768461] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-da0e9b50-a01f-4899-8b09-35884e9fd8ad {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.772874] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec3ac66-df74-4259-a62c-1d766c1479ae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.777992] env[61243]: DEBUG oslo_vmware.api [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for the task: (returnval){ [ 1067.777992] env[61243]: value = "task-1339479" [ 1067.777992] env[61243]: _type = "Task" [ 1067.777992] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.802111] env[61243]: DEBUG oslo_concurrency.lockutils [None req-209aa860-cdb6-4b6d-a15d-e23606ab4f38 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.026s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.808569] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b401cd-79b8-4336-94a6-c700b57eb083 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.815660] env[61243]: DEBUG oslo_vmware.api [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339479, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.834528] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] The volume has not been displaced from its original location: [datastore1] volume-d494ae0c-78b7-4745-85ac-83f826afe53a/volume-d494ae0c-78b7-4745-85ac-83f826afe53a.vmdk. No consolidation needed. {{(pid=61243) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1067.840923] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Reconfiguring VM instance instance-0000005e to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1067.841458] env[61243]: DEBUG oslo_vmware.api [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339472, 'name': PowerOnVM_Task, 'duration_secs': 0.646511} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.845287] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77fd6fcf-6b30-4ce2-945f-830c90ef8ce0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.858719] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1067.858937] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a2aa168a-2ab7-41b4-93cc-d9f0d9fdc54e tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updating instance 'bb7c810f-91d1-43eb-b0cc-fb4c686b1f29' progress to 100 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1067.870496] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339476, 'name': Rename_Task, 'duration_secs': 0.249935} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.871868] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1067.872779] env[61243]: DEBUG oslo_vmware.api [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1067.872779] env[61243]: value = "task-1339481" [ 1067.872779] env[61243]: _type = "Task" [ 1067.872779] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.872779] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-47c9d29e-2733-471b-847f-19a3b86bbf9c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.884912] env[61243]: DEBUG oslo_vmware.api [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339481, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.886229] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1067.886229] env[61243]: value = "task-1339482" [ 1067.886229] env[61243]: _type = "Task" [ 1067.886229] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.897279] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339482, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.955699] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339474, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.576029} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.955997] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] fbf413f0-389c-4924-bf5d-400f95c89a84/fbf413f0-389c-4924-bf5d-400f95c89a84.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1067.956336] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1067.956631] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da596914-aa1d-4eeb-a552-27fef8db440a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.963585] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1067.963585] env[61243]: value = "task-1339483" [ 1067.963585] env[61243]: _type = "Task" [ 1067.963585] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.972400] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339483, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.984296] env[61243]: DEBUG oslo_vmware.api [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339477, 'name': ReconfigVM_Task, 'duration_secs': 0.147387} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.984649] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285783', 'volume_id': 'bce221b9-8c01-45eb-b91e-16b64854ec9c', 'name': 'volume-bce221b9-8c01-45eb-b91e-16b64854ec9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dca3301d-ef3e-48ef-920c-866b4a086ea6', 'attached_at': '', 'detached_at': '', 'volume_id': 'bce221b9-8c01-45eb-b91e-16b64854ec9c', 'serial': 'bce221b9-8c01-45eb-b91e-16b64854ec9c'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1068.031869] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339475, 'name': PowerOnVM_Task, 'duration_secs': 0.696219} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.032237] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1068.032480] env[61243]: INFO nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Took 9.05 seconds to spawn the instance on the hypervisor. [ 1068.032710] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1068.033656] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-010f44b2-9ca6-456a-8892-6fd6a8289063 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.288669] env[61243]: DEBUG oslo_vmware.api [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Task: {'id': task-1339479, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.206285} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.288949] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1068.289166] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1068.289356] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1068.289543] env[61243]: INFO nova.compute.manager [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1068.289786] env[61243]: DEBUG oslo.service.loopingcall [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1068.289981] env[61243]: DEBUG nova.compute.manager [-] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1068.290092] env[61243]: DEBUG nova.network.neutron [-] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1068.386383] env[61243]: DEBUG oslo_vmware.api [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339481, 'name': ReconfigVM_Task, 'duration_secs': 0.336582} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.386753] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Reconfigured VM instance instance-0000005e to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1068.394927] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d748981-11d9-4e3b-984a-4a90c6f66301 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.412445] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339482, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.413825] env[61243]: DEBUG oslo_vmware.api [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1068.413825] env[61243]: value = "task-1339484" [ 1068.413825] env[61243]: _type = "Task" [ 1068.413825] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.422492] env[61243]: DEBUG oslo_vmware.api [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339484, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.474016] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339483, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07367} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.474375] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1068.475282] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa32306c-1997-4a2d-a44e-6c84226346df {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.500482] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] fbf413f0-389c-4924-bf5d-400f95c89a84/fbf413f0-389c-4924-bf5d-400f95c89a84.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1068.500821] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69e427da-0bf8-4720-bda9-228e5e3d35f2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.523088] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1068.523088] env[61243]: value = "task-1339485" [ 1068.523088] env[61243]: _type = "Task" [ 1068.523088] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.539046] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339485, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.540276] env[61243]: DEBUG nova.objects.instance [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lazy-loading 'flavor' on Instance uuid dca3301d-ef3e-48ef-920c-866b4a086ea6 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.556409] env[61243]: INFO nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Took 16.50 seconds to build instance. [ 1068.604298] env[61243]: DEBUG nova.compute.manager [req-0ff05600-c99b-4658-bfd9-0afed8b82839 req-4f713de6-4227-4703-b9b2-27fac5fd3683 service nova] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Received event network-vif-deleted-04ff9536-8994-48d8-aa25-6d4d977da0e6 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1068.604722] env[61243]: INFO nova.compute.manager [req-0ff05600-c99b-4658-bfd9-0afed8b82839 req-4f713de6-4227-4703-b9b2-27fac5fd3683 service nova] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Neutron deleted interface 04ff9536-8994-48d8-aa25-6d4d977da0e6; detaching it from the instance and deleting it from the info cache [ 1068.604926] env[61243]: DEBUG nova.network.neutron [req-0ff05600-c99b-4658-bfd9-0afed8b82839 req-4f713de6-4227-4703-b9b2-27fac5fd3683 service nova] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.666495] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Volume attach. Driver type: vmdk {{(pid=61243) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1068.666872] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285791', 'volume_id': '77ec08cb-5d18-4bee-b234-4aca1c5ce999', 'name': 'volume-77ec08cb-5d18-4bee-b234-4aca1c5ce999', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '60fa2863-f0aa-4798-8e5f-e93df47f93e8', 'attached_at': '', 'detached_at': '', 'volume_id': '77ec08cb-5d18-4bee-b234-4aca1c5ce999', 'serial': '77ec08cb-5d18-4bee-b234-4aca1c5ce999'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1068.668249] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b425b321-23e4-4e48-a23c-234289b1c190 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.695247] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87189a03-8155-4a10-aac9-a01150c1d55c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.722308] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] volume-77ec08cb-5d18-4bee-b234-4aca1c5ce999/volume-77ec08cb-5d18-4bee-b234-4aca1c5ce999.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1068.722629] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0951808a-a922-430b-928a-a5b865fa816b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.743636] env[61243]: DEBUG oslo_vmware.api [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1068.743636] env[61243]: value = "task-1339486" [ 1068.743636] env[61243]: _type = "Task" [ 1068.743636] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.752616] env[61243]: DEBUG oslo_vmware.api [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339486, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.897648] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339482, 'name': PowerOnVM_Task, 'duration_secs': 0.744918} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.897951] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1068.898198] env[61243]: INFO nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Took 7.58 seconds to spawn the instance on the hypervisor. [ 1068.898406] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1068.899234] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9408bb7d-729c-4f2a-8a85-4175edbb21d6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.924162] env[61243]: DEBUG oslo_vmware.api [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339484, 'name': ReconfigVM_Task, 'duration_secs': 0.223341} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.924651] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285780', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'name': 'volume-d494ae0c-78b7-4745-85ac-83f826afe53a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5f1c246d-2732-4da4-a00b-b92e2990aa67', 'attached_at': '', 'detached_at': '', 'volume_id': 'd494ae0c-78b7-4745-85ac-83f826afe53a', 'serial': 'd494ae0c-78b7-4745-85ac-83f826afe53a'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1069.033913] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339485, 'name': ReconfigVM_Task, 'duration_secs': 0.353143} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.036870] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Reconfigured VM instance instance-0000006b to attach disk [datastore1] fbf413f0-389c-4924-bf5d-400f95c89a84/fbf413f0-389c-4924-bf5d-400f95c89a84.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1069.036870] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ffee8a9e-1e82-42ff-bece-3248fcc746db {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.060017] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1069.060017] env[61243]: value = "task-1339487" [ 1069.060017] env[61243]: _type = "Task" [ 1069.060017] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.063266] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "3fabef9e-13ff-41b6-a4eb-ee65b66c37fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.026s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.070258] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339487, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.074684] env[61243]: DEBUG nova.network.neutron [-] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.107866] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eeaf2497-b416-44bd-820f-26af2e29e1ff {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.120576] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f723bf27-710e-45fe-99f9-b96fad8fdf43 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.162541] env[61243]: DEBUG nova.compute.manager [req-0ff05600-c99b-4658-bfd9-0afed8b82839 req-4f713de6-4227-4703-b9b2-27fac5fd3683 service nova] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Detach interface failed, port_id=04ff9536-8994-48d8-aa25-6d4d977da0e6, reason: Instance df141b12-1d34-4a47-b6bb-15cb75d266c3 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1069.254316] env[61243]: DEBUG oslo_vmware.api [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339486, 'name': ReconfigVM_Task, 'duration_secs': 0.404454} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.254621] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Reconfigured VM instance instance-00000062 to attach disk [datastore1] volume-77ec08cb-5d18-4bee-b234-4aca1c5ce999/volume-77ec08cb-5d18-4bee-b234-4aca1c5ce999.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1069.259560] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e1a18c0-0fb4-4420-8a9e-16a44c644624 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.275789] env[61243]: DEBUG oslo_vmware.api [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1069.275789] env[61243]: value = "task-1339488" [ 1069.275789] env[61243]: _type = "Task" [ 1069.275789] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.284847] env[61243]: DEBUG oslo_vmware.api [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339488, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.416141] env[61243]: INFO nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Took 17.36 seconds to build instance. [ 1069.468176] env[61243]: DEBUG nova.objects.instance [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lazy-loading 'flavor' on Instance uuid 5f1c246d-2732-4da4-a00b-b92e2990aa67 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.548360] env[61243]: DEBUG oslo_concurrency.lockutils [None req-729e98a9-a119-47d8-a317-e71c39385bba tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "dca3301d-ef3e-48ef-920c-866b4a086ea6" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.286s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.572111] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339487, 'name': Rename_Task, 'duration_secs': 0.360446} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.572458] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1069.572672] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-005677b7-c883-47d5-80a9-ec63f8b10d50 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.576818] env[61243]: INFO nova.compute.manager [-] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Took 1.29 seconds to deallocate network for instance. [ 1069.584406] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1069.584406] env[61243]: value = "task-1339489" [ 1069.584406] env[61243]: _type = "Task" [ 1069.584406] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.595272] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339489, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.787020] env[61243]: DEBUG oslo_vmware.api [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339488, 'name': ReconfigVM_Task, 'duration_secs': 0.167418} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.787389] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285791', 'volume_id': '77ec08cb-5d18-4bee-b234-4aca1c5ce999', 'name': 'volume-77ec08cb-5d18-4bee-b234-4aca1c5ce999', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '60fa2863-f0aa-4798-8e5f-e93df47f93e8', 'attached_at': '', 'detached_at': '', 'volume_id': '77ec08cb-5d18-4bee-b234-4aca1c5ce999', 'serial': '77ec08cb-5d18-4bee-b234-4aca1c5ce999'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1069.918515] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "26ca8850-4785-4c02-9dd6-3788d0164510" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.879s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.086542] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.086888] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.087177] env[61243]: DEBUG nova.objects.instance [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lazy-loading 'resources' on Instance uuid df141b12-1d34-4a47-b6bb-15cb75d266c3 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.099575] env[61243]: DEBUG oslo_vmware.api [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339489, 'name': PowerOnVM_Task, 'duration_secs': 0.505392} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.100267] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1070.100503] env[61243]: INFO nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Took 6.43 seconds to spawn the instance on the hypervisor. [ 1070.100710] env[61243]: DEBUG nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1070.101589] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9143de8f-05e0-4a7d-9276-bdc02167ab06 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.305634] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "interface-466f6a0a-9bc2-4709-8625-e6fc65b68f80-deadb902-8527-45ad-aa9c-bb805228869f" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.305921] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-466f6a0a-9bc2-4709-8625-e6fc65b68f80-deadb902-8527-45ad-aa9c-bb805228869f" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.306324] env[61243]: DEBUG nova.objects.instance [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lazy-loading 'flavor' on Instance uuid 466f6a0a-9bc2-4709-8625-e6fc65b68f80 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.317700] env[61243]: DEBUG oslo_concurrency.lockutils [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.317944] env[61243]: DEBUG oslo_concurrency.lockutils [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.318156] env[61243]: DEBUG nova.compute.manager [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Going to confirm migration 4 {{(pid=61243) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1070.477958] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d09b3d27-9766-4126-8f76-2e2a6c85f5ad tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "5f1c246d-2732-4da4-a00b-b92e2990aa67" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.277s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.569290] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "dca3301d-ef3e-48ef-920c-866b4a086ea6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.569570] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "dca3301d-ef3e-48ef-920c-866b4a086ea6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.569924] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "dca3301d-ef3e-48ef-920c-866b4a086ea6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.570206] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "dca3301d-ef3e-48ef-920c-866b4a086ea6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.570394] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "dca3301d-ef3e-48ef-920c-866b4a086ea6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.572606] env[61243]: INFO nova.compute.manager [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Terminating instance [ 1070.575134] env[61243]: DEBUG nova.compute.manager [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1070.575334] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1070.576172] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b47b4425-53e7-4926-9954-9f5ecc0e03c8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.585401] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1070.585671] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc7967fa-86da-45f9-8cda-e76800716630 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.595982] env[61243]: DEBUG oslo_vmware.api [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1070.595982] env[61243]: value = "task-1339490" [ 1070.595982] env[61243]: _type = "Task" [ 1070.595982] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.606088] env[61243]: DEBUG oslo_vmware.api [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339490, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.624743] env[61243]: INFO nova.compute.manager [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Took 18.55 seconds to build instance. [ 1070.772713] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7f7e845-df18-480e-b0b7-7b166729f95a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.780475] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd2484a4-7499-447d-943b-fdfa431b3031 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.816685] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9051461e-202e-4efb-a378-a760fdf5e5e4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.828264] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b749ab-bdc7-4dcc-909a-cfdaab6c1679 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.832759] env[61243]: DEBUG nova.objects.instance [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lazy-loading 'flavor' on Instance uuid 60fa2863-f0aa-4798-8e5f-e93df47f93e8 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.846619] env[61243]: DEBUG nova.compute.provider_tree [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.925118] env[61243]: DEBUG oslo_concurrency.lockutils [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "refresh_cache-bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.925118] env[61243]: DEBUG oslo_concurrency.lockutils [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "refresh_cache-bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.925411] env[61243]: DEBUG nova.network.neutron [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1070.925666] env[61243]: DEBUG nova.objects.instance [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lazy-loading 'info_cache' on Instance uuid bb7c810f-91d1-43eb-b0cc-fb4c686b1f29 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.968391] env[61243]: DEBUG nova.objects.instance [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lazy-loading 'pci_requests' on Instance uuid 466f6a0a-9bc2-4709-8625-e6fc65b68f80 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.107663] env[61243]: DEBUG oslo_vmware.api [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339490, 'name': PowerOffVM_Task, 'duration_secs': 0.467982} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.108322] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1071.108528] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1071.108801] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ae5e9a44-bfb6-40ce-8359-3ce8c7233bbf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.127412] env[61243]: DEBUG oslo_concurrency.lockutils [None req-04d0c771-7d8d-4969-837c-cb6eebed1ae5 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "fbf413f0-389c-4924-bf5d-400f95c89a84" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.065s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.186078] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1071.186375] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1071.186689] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Deleting the datastore file [datastore2] dca3301d-ef3e-48ef-920c-866b4a086ea6 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1071.187080] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-60ff3b7d-7b3e-4ccb-b44a-569218eec997 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.194840] env[61243]: DEBUG oslo_vmware.api [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1071.194840] env[61243]: value = "task-1339493" [ 1071.194840] env[61243]: _type = "Task" [ 1071.194840] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.203324] env[61243]: DEBUG oslo_vmware.api [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339493, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.338968] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8ff1d9a7-139f-40d0-b2a9-ba3a5f9a0cb9 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.285s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.350113] env[61243]: DEBUG nova.scheduler.client.report [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.471057] env[61243]: DEBUG nova.objects.base [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Object Instance<466f6a0a-9bc2-4709-8625-e6fc65b68f80> lazy-loaded attributes: flavor,pci_requests {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1071.471057] env[61243]: DEBUG nova.network.neutron [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1071.537228] env[61243]: DEBUG nova.policy [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b32f4788825f4a8e82e028cc10347746', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '683c027e4adc4cc4a6b8747a8c8c0de9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1071.708046] env[61243]: DEBUG oslo_vmware.api [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339493, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.19582} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.708311] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1071.708499] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1071.708680] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1071.708859] env[61243]: INFO nova.compute.manager [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1071.709122] env[61243]: DEBUG oslo.service.loopingcall [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.709441] env[61243]: DEBUG nova.compute.manager [-] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1071.709542] env[61243]: DEBUG nova.network.neutron [-] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1071.856382] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.769s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.869515] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "5f1c246d-2732-4da4-a00b-b92e2990aa67" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.869797] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "5f1c246d-2732-4da4-a00b-b92e2990aa67" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.870031] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "5f1c246d-2732-4da4-a00b-b92e2990aa67-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.870236] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "5f1c246d-2732-4da4-a00b-b92e2990aa67-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.870410] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "5f1c246d-2732-4da4-a00b-b92e2990aa67-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.875660] env[61243]: INFO nova.compute.manager [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Terminating instance [ 1071.881139] env[61243]: INFO nova.scheduler.client.report [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Deleted allocations for instance df141b12-1d34-4a47-b6bb-15cb75d266c3 [ 1071.886385] env[61243]: DEBUG nova.compute.manager [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1071.886385] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1071.886385] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58efbd93-4e0a-49ff-bf2f-24d4771bdcfe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.897619] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1071.897909] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-314b20c6-0199-4af0-aa3e-c9530e003fdc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.905234] env[61243]: DEBUG oslo_vmware.api [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1071.905234] env[61243]: value = "task-1339494" [ 1071.905234] env[61243]: _type = "Task" [ 1071.905234] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.914496] env[61243]: DEBUG oslo_vmware.api [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339494, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.184027] env[61243]: DEBUG nova.compute.manager [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Stashing vm_state: active {{(pid=61243) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1072.397265] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4ce08d7f-2785-4b71-890c-cc102dc9c5af tempest-ServersTestJSON-1786371282 tempest-ServersTestJSON-1786371282-project-member] Lock "df141b12-1d34-4a47-b6bb-15cb75d266c3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.732s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.418403] env[61243]: DEBUG oslo_vmware.api [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339494, 'name': PowerOffVM_Task, 'duration_secs': 0.376848} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.419495] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1072.419495] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1072.419495] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-562453c6-1209-4f54-8d25-6de1761a76bc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.469974] env[61243]: DEBUG nova.network.neutron [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updating instance_info_cache with network_info: [{"id": "a968f6e7-b39a-4dd8-abf8-bf0919565a72", "address": "fa:16:3e:80:14:77", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa968f6e7-b3", "ovs_interfaceid": "a968f6e7-b39a-4dd8-abf8-bf0919565a72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.596097] env[61243]: DEBUG nova.compute.manager [req-d03e3d0a-0156-494a-b954-b6378345294b req-53852658-efca-4ea0-a6e0-cfcbab775102 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Received event network-vif-deleted-ed08f6b9-7027-417b-91d5-e9639b3fed92 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1072.596097] env[61243]: INFO nova.compute.manager [req-d03e3d0a-0156-494a-b954-b6378345294b req-53852658-efca-4ea0-a6e0-cfcbab775102 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Neutron deleted interface ed08f6b9-7027-417b-91d5-e9639b3fed92; detaching it from the instance and deleting it from the info cache [ 1072.596097] env[61243]: DEBUG nova.network.neutron [req-d03e3d0a-0156-494a-b954-b6378345294b req-53852658-efca-4ea0-a6e0-cfcbab775102 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.705669] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.706042] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.850025] env[61243]: DEBUG oslo_concurrency.lockutils [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "3fabef9e-13ff-41b6-a4eb-ee65b66c37fe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.850323] env[61243]: DEBUG oslo_concurrency.lockutils [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "3fabef9e-13ff-41b6-a4eb-ee65b66c37fe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.850540] env[61243]: DEBUG oslo_concurrency.lockutils [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "3fabef9e-13ff-41b6-a4eb-ee65b66c37fe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.850745] env[61243]: DEBUG oslo_concurrency.lockutils [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "3fabef9e-13ff-41b6-a4eb-ee65b66c37fe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.851041] env[61243]: DEBUG oslo_concurrency.lockutils [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "3fabef9e-13ff-41b6-a4eb-ee65b66c37fe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.855721] env[61243]: INFO nova.compute.manager [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Terminating instance [ 1072.857591] env[61243]: DEBUG nova.compute.manager [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1072.857796] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1072.858622] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba537516-4b81-4355-83d0-e5da83211545 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.867706] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1072.867706] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80789943-8141-4214-a9e2-31380cca0d75 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.876802] env[61243]: DEBUG oslo_vmware.api [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1072.876802] env[61243]: value = "task-1339496" [ 1072.876802] env[61243]: _type = "Task" [ 1072.876802] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.889858] env[61243]: DEBUG oslo_vmware.api [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339496, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.975306] env[61243]: DEBUG oslo_concurrency.lockutils [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "refresh_cache-bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.975620] env[61243]: DEBUG nova.objects.instance [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lazy-loading 'migration_context' on Instance uuid bb7c810f-91d1-43eb-b0cc-fb4c686b1f29 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1073.043762] env[61243]: DEBUG nova.network.neutron [-] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.099066] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-95f1bb7d-af04-42eb-b701-43d235923fa8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.110870] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-868bdd16-7027-476f-b42c-d7ff255032e4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.126508] env[61243]: DEBUG nova.network.neutron [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Successfully updated port: deadb902-8527-45ad-aa9c-bb805228869f {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1073.153226] env[61243]: DEBUG nova.compute.manager [req-d03e3d0a-0156-494a-b954-b6378345294b req-53852658-efca-4ea0-a6e0-cfcbab775102 service nova] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Detach interface failed, port_id=ed08f6b9-7027-417b-91d5-e9639b3fed92, reason: Instance dca3301d-ef3e-48ef-920c-866b4a086ea6 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1073.211164] env[61243]: INFO nova.compute.claims [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1073.366689] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1073.367058] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1073.367280] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleting the datastore file [datastore1] 5f1c246d-2732-4da4-a00b-b92e2990aa67 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1073.367559] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cabfa596-c569-4d9a-9fbf-8b4cc6b2f4a8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.376537] env[61243]: DEBUG oslo_vmware.api [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1073.376537] env[61243]: value = "task-1339497" [ 1073.376537] env[61243]: _type = "Task" [ 1073.376537] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.389734] env[61243]: DEBUG oslo_vmware.api [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339496, 'name': PowerOffVM_Task, 'duration_secs': 0.30212} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.392628] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1073.392886] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1073.393176] env[61243]: DEBUG oslo_vmware.api [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339497, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.393399] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fbb32ba4-9fe5-4fc1-86af-f9cb9d4bfb8c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.478682] env[61243]: DEBUG nova.objects.base [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1073.479877] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ba639f-2e00-4398-bc11-c268072baf48 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.502845] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ac38332-b1cd-43b4-9941-d739976ef869 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.512648] env[61243]: DEBUG oslo_vmware.api [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1073.512648] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52fd8583-e8c1-a4c1-c159-f59d45d08059" [ 1073.512648] env[61243]: _type = "Task" [ 1073.512648] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.517613] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1073.517847] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1073.520370] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Deleting the datastore file [datastore1] 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1073.520370] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-24815397-d3bd-4e3d-b22e-137d3056d03e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.525337] env[61243]: DEBUG oslo_vmware.api [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52fd8583-e8c1-a4c1-c159-f59d45d08059, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.531104] env[61243]: DEBUG oslo_vmware.api [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1073.531104] env[61243]: value = "task-1339499" [ 1073.531104] env[61243]: _type = "Task" [ 1073.531104] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.539477] env[61243]: DEBUG oslo_vmware.api [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339499, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.545845] env[61243]: INFO nova.compute.manager [-] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Took 1.84 seconds to deallocate network for instance. [ 1073.630036] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.630036] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.630421] env[61243]: DEBUG nova.network.neutron [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1073.720395] env[61243]: INFO nova.compute.resource_tracker [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating resource usage from migration c6475dac-275e-43dc-b28b-6d7f0683bc1a [ 1073.888359] env[61243]: DEBUG oslo_vmware.api [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339497, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.299191} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.888359] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1073.888359] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1073.888703] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1073.888703] env[61243]: INFO nova.compute.manager [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Took 2.00 seconds to destroy the instance on the hypervisor. [ 1073.888978] env[61243]: DEBUG oslo.service.loopingcall [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1073.889610] env[61243]: DEBUG nova.compute.manager [-] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1073.889610] env[61243]: DEBUG nova.network.neutron [-] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1073.900244] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3469d82a-b7c9-4270-adf2-7efac4eb07fd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.908228] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7eb9a8b-beb4-4d26-941f-01e81dcb250b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.941584] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd66f7c5-e478-461c-80b1-4a7cf951473f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.951599] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c42b2f-4349-4c26-a915-414cc6a4a0b2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.968763] env[61243]: DEBUG nova.compute.provider_tree [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1074.024839] env[61243]: DEBUG oslo_vmware.api [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52fd8583-e8c1-a4c1-c159-f59d45d08059, 'name': SearchDatastore_Task, 'duration_secs': 0.022329} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.025193] env[61243]: DEBUG oslo_concurrency.lockutils [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.040232] env[61243]: DEBUG oslo_vmware.api [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339499, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.352246} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.040521] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1074.040719] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1074.040914] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1074.041186] env[61243]: INFO nova.compute.manager [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1074.041356] env[61243]: DEBUG oslo.service.loopingcall [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1074.041562] env[61243]: DEBUG nova.compute.manager [-] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1074.041629] env[61243]: DEBUG nova.network.neutron [-] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1074.051675] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.213697] env[61243]: WARNING nova.network.neutron [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] 7724b87f-3721-42a5-81ce-27ae98fc99c2 already exists in list: networks containing: ['7724b87f-3721-42a5-81ce-27ae98fc99c2']. ignoring it [ 1074.494269] env[61243]: ERROR nova.scheduler.client.report [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [req-f0dad803-374f-464a-a73a-1da5b8098103] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8f3c282f-58fe-4c5d-80db-5a142cf023e1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f0dad803-374f-464a-a73a-1da5b8098103"}]} [ 1074.513476] env[61243]: DEBUG nova.scheduler.client.report [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Refreshing inventories for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1074.535177] env[61243]: DEBUG nova.scheduler.client.report [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Updating ProviderTree inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1074.535177] env[61243]: DEBUG nova.compute.provider_tree [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1074.546785] env[61243]: DEBUG nova.scheduler.client.report [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Refreshing aggregate associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, aggregates: None {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1074.564842] env[61243]: DEBUG nova.scheduler.client.report [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Refreshing trait associations for resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61243) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1074.648986] env[61243]: DEBUG nova.compute.manager [req-0cec4d9f-c8c9-43fc-afb1-33f449eda313 req-31b093f6-74a1-43e9-92e1-18dbb2e3bc4e service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Received event network-vif-plugged-deadb902-8527-45ad-aa9c-bb805228869f {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1074.649255] env[61243]: DEBUG oslo_concurrency.lockutils [req-0cec4d9f-c8c9-43fc-afb1-33f449eda313 req-31b093f6-74a1-43e9-92e1-18dbb2e3bc4e service nova] Acquiring lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.650053] env[61243]: DEBUG oslo_concurrency.lockutils [req-0cec4d9f-c8c9-43fc-afb1-33f449eda313 req-31b093f6-74a1-43e9-92e1-18dbb2e3bc4e service nova] Lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.650347] env[61243]: DEBUG oslo_concurrency.lockutils [req-0cec4d9f-c8c9-43fc-afb1-33f449eda313 req-31b093f6-74a1-43e9-92e1-18dbb2e3bc4e service nova] Lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.650783] env[61243]: DEBUG nova.compute.manager [req-0cec4d9f-c8c9-43fc-afb1-33f449eda313 req-31b093f6-74a1-43e9-92e1-18dbb2e3bc4e service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] No waiting events found dispatching network-vif-plugged-deadb902-8527-45ad-aa9c-bb805228869f {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1074.652720] env[61243]: WARNING nova.compute.manager [req-0cec4d9f-c8c9-43fc-afb1-33f449eda313 req-31b093f6-74a1-43e9-92e1-18dbb2e3bc4e service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Received unexpected event network-vif-plugged-deadb902-8527-45ad-aa9c-bb805228869f for instance with vm_state active and task_state None. [ 1074.652720] env[61243]: DEBUG nova.compute.manager [req-0cec4d9f-c8c9-43fc-afb1-33f449eda313 req-31b093f6-74a1-43e9-92e1-18dbb2e3bc4e service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Received event network-changed-deadb902-8527-45ad-aa9c-bb805228869f {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1074.652720] env[61243]: DEBUG nova.compute.manager [req-0cec4d9f-c8c9-43fc-afb1-33f449eda313 req-31b093f6-74a1-43e9-92e1-18dbb2e3bc4e service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Refreshing instance network info cache due to event network-changed-deadb902-8527-45ad-aa9c-bb805228869f. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1074.652720] env[61243]: DEBUG oslo_concurrency.lockutils [req-0cec4d9f-c8c9-43fc-afb1-33f449eda313 req-31b093f6-74a1-43e9-92e1-18dbb2e3bc4e service nova] Acquiring lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.773941] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13f04b0-1f84-440f-882b-976ae41cea6c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.782267] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98f2825f-0e28-4a77-bf9c-25b02a5a5230 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.829548] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2114d447-af1d-4954-a0bd-53a84d528e3d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.835566] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e28b9d29-e22b-4c95-be1e-083f77368353 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.841259] env[61243]: DEBUG nova.network.neutron [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Updating instance_info_cache with network_info: [{"id": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "address": "fa:16:3e:43:9a:6f", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde6bbdd5-c1", "ovs_interfaceid": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "deadb902-8527-45ad-aa9c-bb805228869f", "address": "fa:16:3e:17:a0:99", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdeadb902-85", "ovs_interfaceid": "deadb902-8527-45ad-aa9c-bb805228869f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.855264] env[61243]: DEBUG nova.compute.provider_tree [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1075.161885] env[61243]: DEBUG nova.network.neutron [-] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.330735] env[61243]: DEBUG nova.network.neutron [-] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.344030] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.344678] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.344845] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.345441] env[61243]: DEBUG oslo_concurrency.lockutils [req-0cec4d9f-c8c9-43fc-afb1-33f449eda313 req-31b093f6-74a1-43e9-92e1-18dbb2e3bc4e service nova] Acquired lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.345634] env[61243]: DEBUG nova.network.neutron [req-0cec4d9f-c8c9-43fc-afb1-33f449eda313 req-31b093f6-74a1-43e9-92e1-18dbb2e3bc4e service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Refreshing network info cache for port deadb902-8527-45ad-aa9c-bb805228869f {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1075.347407] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6cb525-1f53-422b-b7aa-a594bd0d8c88 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.370651] env[61243]: DEBUG nova.virt.hardware [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1075.370784] env[61243]: DEBUG nova.virt.hardware [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1075.370892] env[61243]: DEBUG nova.virt.hardware [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1075.371094] env[61243]: DEBUG nova.virt.hardware [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1075.371250] env[61243]: DEBUG nova.virt.hardware [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1075.371398] env[61243]: DEBUG nova.virt.hardware [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1075.371595] env[61243]: DEBUG nova.virt.hardware [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1075.371756] env[61243]: DEBUG nova.virt.hardware [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1075.371921] env[61243]: DEBUG nova.virt.hardware [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1075.372098] env[61243]: DEBUG nova.virt.hardware [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1075.372278] env[61243]: DEBUG nova.virt.hardware [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1075.378908] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Reconfiguring VM to attach interface {{(pid=61243) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1075.379721] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-780e7e86-3253-41e2-81a2-3a515643f417 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.398960] env[61243]: DEBUG oslo_vmware.api [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1075.398960] env[61243]: value = "task-1339501" [ 1075.398960] env[61243]: _type = "Task" [ 1075.398960] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.409814] env[61243]: DEBUG oslo_vmware.api [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339501, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.410732] env[61243]: DEBUG nova.scheduler.client.report [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Updated inventory for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with generation 156 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1075.410940] env[61243]: DEBUG nova.compute.provider_tree [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Updating resource provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 generation from 156 to 157 during operation: update_inventory {{(pid=61243) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1075.411144] env[61243]: DEBUG nova.compute.provider_tree [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Updating inventory in ProviderTree for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1075.664387] env[61243]: INFO nova.compute.manager [-] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Took 1.62 seconds to deallocate network for instance. [ 1075.833789] env[61243]: INFO nova.compute.manager [-] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Took 1.94 seconds to deallocate network for instance. [ 1075.914220] env[61243]: DEBUG oslo_vmware.api [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339501, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.916370] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 3.210s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.916642] env[61243]: INFO nova.compute.manager [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Migrating [ 1075.926454] env[61243]: DEBUG oslo_concurrency.lockutils [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.901s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.079121] env[61243]: DEBUG nova.network.neutron [req-0cec4d9f-c8c9-43fc-afb1-33f449eda313 req-31b093f6-74a1-43e9-92e1-18dbb2e3bc4e service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Updated VIF entry in instance network info cache for port deadb902-8527-45ad-aa9c-bb805228869f. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1076.079477] env[61243]: DEBUG nova.network.neutron [req-0cec4d9f-c8c9-43fc-afb1-33f449eda313 req-31b093f6-74a1-43e9-92e1-18dbb2e3bc4e service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Updating instance_info_cache with network_info: [{"id": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "address": "fa:16:3e:43:9a:6f", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde6bbdd5-c1", "ovs_interfaceid": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "deadb902-8527-45ad-aa9c-bb805228869f", "address": "fa:16:3e:17:a0:99", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdeadb902-85", "ovs_interfaceid": "deadb902-8527-45ad-aa9c-bb805228869f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.171908] env[61243]: DEBUG oslo_concurrency.lockutils [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.302705] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.303123] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.343497] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.381834] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquiring lock "6786bf24-f760-4f9b-947a-25d19ae08253" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.382120] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Lock "6786bf24-f760-4f9b-947a-25d19ae08253" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.412232] env[61243]: DEBUG oslo_vmware.api [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339501, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.440043] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.440306] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.440520] env[61243]: DEBUG nova.network.neutron [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1076.582043] env[61243]: DEBUG oslo_concurrency.lockutils [req-0cec4d9f-c8c9-43fc-afb1-33f449eda313 req-31b093f6-74a1-43e9-92e1-18dbb2e3bc4e service nova] Releasing lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.620375] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4cf543f-7bbe-4512-a68b-123d904229bc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.629461] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea569360-5b08-447b-886d-d44c889923e0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.661148] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893780b0-9621-4145-b6a9-304a1e7037e0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.669726] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a6117e-b889-4f24-bb3d-951d333c0692 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.683962] env[61243]: DEBUG nova.compute.provider_tree [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1076.687498] env[61243]: DEBUG nova.compute.manager [req-16db2b8b-2e35-410c-a557-13a8e098695d req-27b824bb-b8fb-4455-9ee5-1c790ead0723 service nova] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Received event network-vif-deleted-f8c67797-fbfc-41d3-b441-9c5274d8c188 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1076.687704] env[61243]: DEBUG nova.compute.manager [req-16db2b8b-2e35-410c-a557-13a8e098695d req-27b824bb-b8fb-4455-9ee5-1c790ead0723 service nova] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Received event network-vif-deleted-6fddda84-0767-4709-ac83-fc4e6060da3c {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1076.810585] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.810585] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.810811] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.811196] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.811196] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager.update_available_resource {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.884847] env[61243]: DEBUG nova.compute.manager [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1076.909795] env[61243]: DEBUG oslo_vmware.api [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339501, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.194953] env[61243]: DEBUG nova.scheduler.client.report [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1077.276126] env[61243]: DEBUG nova.network.neutron [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating instance_info_cache with network_info: [{"id": "f2a48425-d436-40c9-a2dc-694e6313da6d", "address": "fa:16:3e:34:84:e2", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2a48425-d4", "ovs_interfaceid": "f2a48425-d436-40c9-a2dc-694e6313da6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.314292] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.410900] env[61243]: DEBUG oslo_vmware.api [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339501, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.412884] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.779760] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.914224] env[61243]: DEBUG oslo_vmware.api [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339501, 'name': ReconfigVM_Task, 'duration_secs': 2.038797} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.914224] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.914224] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Reconfigured VM to attach interface {{(pid=61243) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1078.205486] env[61243]: DEBUG oslo_concurrency.lockutils [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.279s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.208384] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.157s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.208587] env[61243]: DEBUG nova.objects.instance [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lazy-loading 'resources' on Instance uuid dca3301d-ef3e-48ef-920c-866b4a086ea6 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1078.418040] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8f6c7263-2125-4849-8fbc-6ca2612f9ea5 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-466f6a0a-9bc2-4709-8625-e6fc65b68f80-deadb902-8527-45ad-aa9c-bb805228869f" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.112s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.878473] env[61243]: INFO nova.scheduler.client.report [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleted allocation for migration 0fcac3af-14cd-43cd-99d0-3e6f3647002e [ 1078.978227] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab560fd0-dc06-450b-a35e-eb5d57ffd2bf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.987269] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fed9ad3-414c-4421-8ea4-4042c093840c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.018182] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a86b303-8b84-45d5-97f1-d684b4bd50e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.025142] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e74f048-e1ed-41f2-b272-d95b521a55ad {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.038019] env[61243]: DEBUG nova.compute.provider_tree [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.297118] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f36d180a-cf42-491e-ba3b-9cb6c70c0f9c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.319557] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating instance '60fa2863-f0aa-4798-8e5f-e93df47f93e8' progress to 0 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1079.384399] env[61243]: DEBUG oslo_concurrency.lockutils [None req-526f41a8-ce57-43f9-a823-40f5c4cc4480 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.066s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.541304] env[61243]: DEBUG nova.scheduler.client.report [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.825624] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1079.825976] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58721cc2-2ae5-4e25-8f2e-6f79a67c0526 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.834299] env[61243]: DEBUG oslo_vmware.api [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1079.834299] env[61243]: value = "task-1339502" [ 1079.834299] env[61243]: _type = "Task" [ 1079.834299] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.845169] env[61243]: DEBUG oslo_vmware.api [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339502, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.047564] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.839s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.050287] env[61243]: DEBUG oslo_concurrency.lockutils [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.878s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.050566] env[61243]: DEBUG nova.objects.instance [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lazy-loading 'resources' on Instance uuid 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1080.068618] env[61243]: INFO nova.scheduler.client.report [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Deleted allocations for instance dca3301d-ef3e-48ef-920c-866b4a086ea6 [ 1080.138961] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "interface-466f6a0a-9bc2-4709-8625-e6fc65b68f80-deadb902-8527-45ad-aa9c-bb805228869f" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.139306] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-466f6a0a-9bc2-4709-8625-e6fc65b68f80-deadb902-8527-45ad-aa9c-bb805228869f" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.140787] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.141019] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.141254] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.141481] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.141657] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.143630] env[61243]: INFO nova.compute.manager [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Terminating instance [ 1080.145407] env[61243]: DEBUG nova.compute.manager [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1080.145601] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1080.146690] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212b0e4c-53a5-496b-b284-e12122922dac {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.155975] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1080.156258] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e065f68-c9aa-45dc-8490-5c63f2024c86 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.163941] env[61243]: DEBUG oslo_vmware.api [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1080.163941] env[61243]: value = "task-1339503" [ 1080.163941] env[61243]: _type = "Task" [ 1080.163941] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.172926] env[61243]: DEBUG oslo_vmware.api [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339503, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.344919] env[61243]: DEBUG oslo_vmware.api [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339502, 'name': PowerOffVM_Task, 'duration_secs': 0.286772} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.345234] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1080.345442] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating instance '60fa2863-f0aa-4798-8e5f-e93df47f93e8' progress to 17 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1080.580210] env[61243]: DEBUG oslo_concurrency.lockutils [None req-10779988-c4fd-4876-956c-cf8b8f6560ad tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "dca3301d-ef3e-48ef-920c-866b4a086ea6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.010s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.606128] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "3d6acdec-ec3a-46b6-858a-135a734ea8ae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.606435] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "3d6acdec-ec3a-46b6-858a-135a734ea8ae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.644316] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.644517] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.645461] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-412c347a-75de-4a56-afab-fbbf157027dd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.673064] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a525ee-c690-4039-b31f-c60aa73537b8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.682107] env[61243]: DEBUG oslo_vmware.api [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339503, 'name': PowerOffVM_Task, 'duration_secs': 0.295536} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.699326] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1080.699542] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1080.705646] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Reconfiguring VM to detach interface {{(pid=61243) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1080.708409] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0bc694f8-3aac-42b2-afa6-d19d562e1398 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.709943] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e373c1e-31ca-45e0-9dbe-87cc2178a5cd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.729876] env[61243]: DEBUG oslo_vmware.api [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1080.729876] env[61243]: value = "task-1339505" [ 1080.729876] env[61243]: _type = "Task" [ 1080.729876] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.741252] env[61243]: DEBUG oslo_vmware.api [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339505, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.781950] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e7a2199-61bd-48e6-96b0-b78bf80e1c2d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.790335] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1080.790595] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1080.790865] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleting the datastore file [datastore2] bb7c810f-91d1-43eb-b0cc-fb4c686b1f29 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1080.793363] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b5524cc3-ec9f-48c0-9e7c-eaaa28b59ae6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.796508] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfac702b-e22b-4298-8bd3-31ce72750347 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.827608] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3b2ed8-648a-428b-a597-198b00e283a5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.830255] env[61243]: DEBUG oslo_vmware.api [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1080.830255] env[61243]: value = "task-1339506" [ 1080.830255] env[61243]: _type = "Task" [ 1080.830255] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.836666] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d202e975-14ba-4460-a735-2b726f668963 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.843440] env[61243]: DEBUG oslo_vmware.api [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339506, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.855166] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1080.855419] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1080.855575] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1080.855767] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1080.855990] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1080.856093] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1080.856308] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1080.856566] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1080.856745] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1080.856920] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1080.857114] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1080.862372] env[61243]: DEBUG nova.compute.provider_tree [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1080.864420] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d6ff0f0-a09b-456f-804d-f1d0a27de58c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.883721] env[61243]: DEBUG oslo_vmware.api [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1080.883721] env[61243]: value = "task-1339507" [ 1080.883721] env[61243]: _type = "Task" [ 1080.883721] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.897924] env[61243]: DEBUG oslo_vmware.api [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339507, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.109301] env[61243]: DEBUG nova.compute.manager [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1081.239974] env[61243]: DEBUG oslo_vmware.api [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.341158] env[61243]: DEBUG oslo_vmware.api [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339506, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.320206} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.341431] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1081.341624] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1081.341809] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1081.341991] env[61243]: INFO nova.compute.manager [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1081.342322] env[61243]: DEBUG oslo.service.loopingcall [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1081.342495] env[61243]: DEBUG nova.compute.manager [-] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1081.342592] env[61243]: DEBUG nova.network.neutron [-] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1081.365445] env[61243]: DEBUG nova.scheduler.client.report [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1081.394068] env[61243]: DEBUG oslo_vmware.api [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339507, 'name': ReconfigVM_Task, 'duration_secs': 0.271716} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.394399] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating instance '60fa2863-f0aa-4798-8e5f-e93df47f93e8' progress to 33 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1081.637138] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.746551] env[61243]: DEBUG oslo_vmware.api [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.766873] env[61243]: DEBUG nova.compute.manager [req-35eacf43-acf0-4dc7-836b-23f0078f0d6c req-f6bea5d2-ef46-4922-a8bf-48111822dca3 service nova] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Received event network-vif-deleted-a968f6e7-b39a-4dd8-abf8-bf0919565a72 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1081.766873] env[61243]: INFO nova.compute.manager [req-35eacf43-acf0-4dc7-836b-23f0078f0d6c req-f6bea5d2-ef46-4922-a8bf-48111822dca3 service nova] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Neutron deleted interface a968f6e7-b39a-4dd8-abf8-bf0919565a72; detaching it from the instance and deleting it from the info cache [ 1081.766873] env[61243]: DEBUG nova.network.neutron [req-35eacf43-acf0-4dc7-836b-23f0078f0d6c req-f6bea5d2-ef46-4922-a8bf-48111822dca3 service nova] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.870594] env[61243]: DEBUG oslo_concurrency.lockutils [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.820s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.873016] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.530s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.873305] env[61243]: DEBUG nova.objects.instance [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lazy-loading 'resources' on Instance uuid 5f1c246d-2732-4da4-a00b-b92e2990aa67 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.892397] env[61243]: INFO nova.scheduler.client.report [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Deleted allocations for instance 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe [ 1081.899826] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1081.900074] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1081.900245] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1081.900431] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1081.900584] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1081.900762] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1081.900933] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1081.901140] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1081.901285] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1081.901452] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1081.901631] env[61243]: DEBUG nova.virt.hardware [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1081.906997] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Reconfiguring VM instance instance-00000062 to detach disk 2000 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1081.907314] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6361e6bb-be3a-47c7-9524-ba391493a8ed {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.927398] env[61243]: DEBUG oslo_vmware.api [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1081.927398] env[61243]: value = "task-1339508" [ 1081.927398] env[61243]: _type = "Task" [ 1081.927398] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.936101] env[61243]: DEBUG oslo_vmware.api [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339508, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.241295] env[61243]: DEBUG nova.network.neutron [-] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.242608] env[61243]: DEBUG oslo_vmware.api [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.268731] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8e3e6bfa-add9-43a0-814e-89a4089c79b0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.278358] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7a7e1ce-0a0a-45d9-a5b2-7a3ed34810f3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.308290] env[61243]: DEBUG nova.compute.manager [req-35eacf43-acf0-4dc7-836b-23f0078f0d6c req-f6bea5d2-ef46-4922-a8bf-48111822dca3 service nova] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Detach interface failed, port_id=a968f6e7-b39a-4dd8-abf8-bf0919565a72, reason: Instance bb7c810f-91d1-43eb-b0cc-fb4c686b1f29 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1082.400307] env[61243]: DEBUG oslo_concurrency.lockutils [None req-51f249ba-00c2-45ec-b62a-583eefadc3c7 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "3fabef9e-13ff-41b6-a4eb-ee65b66c37fe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.549s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.438315] env[61243]: DEBUG oslo_vmware.api [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339508, 'name': ReconfigVM_Task, 'duration_secs': 0.23622} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.438590] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Reconfigured VM instance instance-00000062 to detach disk 2000 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1082.439366] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09adddcd-b790-4938-89e3-768aed83de0c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.463432] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 60fa2863-f0aa-4798-8e5f-e93df47f93e8/60fa2863-f0aa-4798-8e5f-e93df47f93e8.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1082.465635] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c009089c-66bd-42a2-bc00-75e4e08a5909 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.484052] env[61243]: DEBUG oslo_vmware.api [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1082.484052] env[61243]: value = "task-1339509" [ 1082.484052] env[61243]: _type = "Task" [ 1082.484052] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.492299] env[61243]: DEBUG oslo_vmware.api [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339509, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.582966] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ee3aa72-f4a1-413c-84a5-6fd64c3f5078 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.596265] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3691d519-beed-408e-8fc0-1878c502c680 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.627924] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e27422-411c-4914-88cb-92d874bb80f7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.636953] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b76925-7235-4417-8cdd-21abd706a7a1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.651490] env[61243]: DEBUG nova.compute.provider_tree [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1082.743741] env[61243]: INFO nova.compute.manager [-] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Took 1.40 seconds to deallocate network for instance. [ 1082.744104] env[61243]: DEBUG oslo_vmware.api [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.820917] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "d24f3683-169b-46a8-9387-b6fc88b654bd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.821194] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d24f3683-169b-46a8-9387-b6fc88b654bd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.994717] env[61243]: DEBUG oslo_vmware.api [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339509, 'name': ReconfigVM_Task, 'duration_secs': 0.341501} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.995011] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 60fa2863-f0aa-4798-8e5f-e93df47f93e8/60fa2863-f0aa-4798-8e5f-e93df47f93e8.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1082.995337] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating instance '60fa2863-f0aa-4798-8e5f-e93df47f93e8' progress to 50 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1083.154258] env[61243]: DEBUG nova.scheduler.client.report [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1083.245796] env[61243]: DEBUG oslo_vmware.api [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.250996] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.323785] env[61243]: DEBUG nova.compute.manager [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1083.503472] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2817a2d1-c5c0-462f-bac5-7a193a686b57 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.531588] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e273a742-0c2e-4312-9343-d668ea01a755 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.552709] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating instance '60fa2863-f0aa-4798-8e5f-e93df47f93e8' progress to 67 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1083.661643] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.789s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.664210] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.350s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.664504] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.664949] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61243) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1083.665234] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.252s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.667962] env[61243]: INFO nova.compute.claims [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1083.672126] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27381bf0-82aa-49f8-a41c-1cea2a48c7c8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.680712] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7dce660-3327-4046-90b7-0ed9f6c4854f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.686614] env[61243]: INFO nova.scheduler.client.report [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleted allocations for instance 5f1c246d-2732-4da4-a00b-b92e2990aa67 [ 1083.697434] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ccf825-e86c-4364-84d1-c9e6063f2762 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.706315] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e9c8188-5229-4c8b-b95c-cacd1be36373 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.738432] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180571MB free_disk=154GB free_vcpus=48 pci_devices=None {{(pid=61243) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1083.738601] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.748705] env[61243]: DEBUG oslo_vmware.api [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.840988] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.881586] env[61243]: DEBUG oslo_concurrency.lockutils [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "26ca8850-4785-4c02-9dd6-3788d0164510" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.882783] env[61243]: DEBUG oslo_concurrency.lockutils [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "26ca8850-4785-4c02-9dd6-3788d0164510" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.882783] env[61243]: DEBUG oslo_concurrency.lockutils [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "26ca8850-4785-4c02-9dd6-3788d0164510-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.882783] env[61243]: DEBUG oslo_concurrency.lockutils [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "26ca8850-4785-4c02-9dd6-3788d0164510-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.882783] env[61243]: DEBUG oslo_concurrency.lockutils [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "26ca8850-4785-4c02-9dd6-3788d0164510-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.884422] env[61243]: INFO nova.compute.manager [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Terminating instance [ 1083.886308] env[61243]: DEBUG nova.compute.manager [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1083.886520] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1083.887397] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a16bcac9-5077-4006-8e93-33e1a93c40ea {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.895772] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1083.896033] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb29d3bc-1684-4db4-844e-f95a8fa3ca85 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.902281] env[61243]: DEBUG oslo_vmware.api [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1083.902281] env[61243]: value = "task-1339510" [ 1083.902281] env[61243]: _type = "Task" [ 1083.902281] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.911331] env[61243]: DEBUG oslo_vmware.api [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339510, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.945230] env[61243]: DEBUG oslo_concurrency.lockutils [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "fbf413f0-389c-4924-bf5d-400f95c89a84" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.945496] env[61243]: DEBUG oslo_concurrency.lockutils [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "fbf413f0-389c-4924-bf5d-400f95c89a84" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.945718] env[61243]: DEBUG oslo_concurrency.lockutils [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "fbf413f0-389c-4924-bf5d-400f95c89a84-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.945930] env[61243]: DEBUG oslo_concurrency.lockutils [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "fbf413f0-389c-4924-bf5d-400f95c89a84-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.946149] env[61243]: DEBUG oslo_concurrency.lockutils [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "fbf413f0-389c-4924-bf5d-400f95c89a84-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.949368] env[61243]: INFO nova.compute.manager [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Terminating instance [ 1083.951511] env[61243]: DEBUG nova.compute.manager [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1083.951718] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1083.952727] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe7bfb10-9733-4393-9fa6-f450f1ccb87f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.960926] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1083.961197] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8577684-ae5f-4b63-baf2-7b89c8468651 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.968340] env[61243]: DEBUG oslo_vmware.api [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1083.968340] env[61243]: value = "task-1339511" [ 1083.968340] env[61243]: _type = "Task" [ 1083.968340] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.978111] env[61243]: DEBUG oslo_vmware.api [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339511, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.205796] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ad8abe4c-c789-4fd8-9d6d-335fde3cfe40 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "5f1c246d-2732-4da4-a00b-b92e2990aa67" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.336s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.206815] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "486ebf65-4376-424c-9deb-f1ef70e44be3" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.207192] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.207269] env[61243]: INFO nova.compute.manager [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Shelving [ 1084.249876] env[61243]: DEBUG oslo_vmware.api [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.413163] env[61243]: DEBUG oslo_vmware.api [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339510, 'name': PowerOffVM_Task, 'duration_secs': 0.214507} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.413459] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1084.413688] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1084.413943] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ae3e743c-12c6-4398-84d7-c8fa9aff66c4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.478509] env[61243]: DEBUG oslo_vmware.api [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339511, 'name': PowerOffVM_Task, 'duration_secs': 0.196347} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.479208] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1084.479511] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1084.479682] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e1e09ec-7c84-4c05-998b-b28cb3384473 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.482375] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1084.482628] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1084.482844] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Deleting the datastore file [datastore1] 26ca8850-4785-4c02-9dd6-3788d0164510 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1084.483459] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6672c99e-3986-4d0d-9e1a-131b488b348a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.491569] env[61243]: DEBUG oslo_vmware.api [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1084.491569] env[61243]: value = "task-1339514" [ 1084.491569] env[61243]: _type = "Task" [ 1084.491569] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.499587] env[61243]: DEBUG oslo_vmware.api [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339514, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.555016] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1084.555407] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1084.555759] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Deleting the datastore file [datastore1] fbf413f0-389c-4924-bf5d-400f95c89a84 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1084.556104] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f4f8dc2-a858-48b1-9b00-f423f238b7ea {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.565075] env[61243]: DEBUG oslo_vmware.api [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for the task: (returnval){ [ 1084.565075] env[61243]: value = "task-1339515" [ 1084.565075] env[61243]: _type = "Task" [ 1084.565075] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.575754] env[61243]: DEBUG oslo_vmware.api [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339515, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.722636] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1084.723284] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a07bab8c-6874-453f-99c2-80c4d76670a6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.731402] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1084.731402] env[61243]: value = "task-1339516" [ 1084.731402] env[61243]: _type = "Task" [ 1084.731402] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.747213] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339516, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.755610] env[61243]: DEBUG oslo_vmware.api [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.855146] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c8d52a8-0298-40ec-96e3-ac2aa72f5eca {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.863319] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d096ea83-587f-4529-a63a-6818f3d7aca4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.894282] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b58b88-2366-47db-8c62-371095bf21ef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.903415] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9cdaee-c8a7-46af-8e59-eb8d9bd54c53 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.917618] env[61243]: DEBUG nova.compute.provider_tree [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1085.003231] env[61243]: DEBUG oslo_vmware.api [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339514, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166726} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.003514] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1085.003737] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1085.004012] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1085.004257] env[61243]: INFO nova.compute.manager [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1085.004516] env[61243]: DEBUG oslo.service.loopingcall [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1085.004710] env[61243]: DEBUG nova.compute.manager [-] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1085.004809] env[61243]: DEBUG nova.network.neutron [-] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1085.074805] env[61243]: DEBUG oslo_vmware.api [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Task: {'id': task-1339515, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166753} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.075098] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1085.075297] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1085.075613] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1085.075680] env[61243]: INFO nova.compute.manager [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1085.075903] env[61243]: DEBUG oslo.service.loopingcall [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1085.076113] env[61243]: DEBUG nova.compute.manager [-] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1085.076211] env[61243]: DEBUG nova.network.neutron [-] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1085.243793] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339516, 'name': PowerOffVM_Task, 'duration_secs': 0.251274} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.247102] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1085.247828] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb8779a-12ea-4e1f-bb08-cf16009c6f4a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.250588] env[61243]: DEBUG nova.network.neutron [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Port f2a48425-d436-40c9-a2dc-694e6313da6d binding to destination host cpu-1 is already ACTIVE {{(pid=61243) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1085.257276] env[61243]: DEBUG oslo_vmware.api [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.271506] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee62ecc0-356b-4046-bf08-bf1173eab1cb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.421753] env[61243]: DEBUG nova.scheduler.client.report [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1085.433097] env[61243]: DEBUG nova.compute.manager [req-743447d6-cbe8-42e7-b1f1-67e5eaae9ed0 req-68b902f3-5c95-4c08-be21-738456816fe3 service nova] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Received event network-vif-deleted-78eac2ea-b988-4138-a10a-a0fd00a3cdb1 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1085.433318] env[61243]: INFO nova.compute.manager [req-743447d6-cbe8-42e7-b1f1-67e5eaae9ed0 req-68b902f3-5c95-4c08-be21-738456816fe3 service nova] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Neutron deleted interface 78eac2ea-b988-4138-a10a-a0fd00a3cdb1; detaching it from the instance and deleting it from the info cache [ 1085.433501] env[61243]: DEBUG nova.network.neutron [req-743447d6-cbe8-42e7-b1f1-67e5eaae9ed0 req-68b902f3-5c95-4c08-be21-738456816fe3 service nova] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.534454] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "91b14a22-811a-4151-9769-9c684ec993db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.534692] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "91b14a22-811a-4151-9769-9c684ec993db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.757453] env[61243]: DEBUG oslo_vmware.api [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.783032] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Creating Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1085.783188] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b464f2e2-6419-45ee-94ad-75bd3c0cb176 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.792819] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1085.792819] env[61243]: value = "task-1339517" [ 1085.792819] env[61243]: _type = "Task" [ 1085.792819] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.802546] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339517, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.912034] env[61243]: DEBUG nova.network.neutron [-] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.927710] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.262s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.928312] env[61243]: DEBUG nova.compute.manager [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1085.930581] env[61243]: DEBUG nova.network.neutron [-] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.931775] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.295s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.933534] env[61243]: INFO nova.compute.claims [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1085.937147] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f67edc3-cf53-4979-99ec-11167590d5e0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.949842] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66214b12-931a-43c0-893a-0f05ba14cafb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.981552] env[61243]: DEBUG nova.compute.manager [req-743447d6-cbe8-42e7-b1f1-67e5eaae9ed0 req-68b902f3-5c95-4c08-be21-738456816fe3 service nova] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Detach interface failed, port_id=78eac2ea-b988-4138-a10a-a0fd00a3cdb1, reason: Instance 26ca8850-4785-4c02-9dd6-3788d0164510 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1086.036571] env[61243]: DEBUG nova.compute.manager [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1086.254577] env[61243]: DEBUG oslo_vmware.api [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339505, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.271500] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.271731] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.271911] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.303296] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339517, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.415013] env[61243]: INFO nova.compute.manager [-] [instance: 26ca8850-4785-4c02-9dd6-3788d0164510] Took 1.41 seconds to deallocate network for instance. [ 1086.433945] env[61243]: DEBUG nova.compute.utils [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1086.436015] env[61243]: DEBUG nova.compute.manager [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Not allocating networking since 'none' was specified. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1086.439912] env[61243]: INFO nova.compute.manager [-] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Took 1.36 seconds to deallocate network for instance. [ 1086.559909] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.755550] env[61243]: DEBUG oslo_vmware.api [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339505, 'name': ReconfigVM_Task, 'duration_secs': 5.790916} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.756104] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.756218] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Reconfigured VM to detach interface {{(pid=61243) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1086.804809] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339517, 'name': CreateSnapshot_Task, 'duration_secs': 0.745693} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.804809] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Created Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1086.805383] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc71678-4d47-4fd2-a99c-1562777845cf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.922945] env[61243]: DEBUG oslo_concurrency.lockutils [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.940708] env[61243]: DEBUG nova.compute.manager [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1086.948887] env[61243]: DEBUG oslo_concurrency.lockutils [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.117708] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-062e1c54-ff3e-43b5-9a43-f0cbc0e2516b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.127936] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c4d34d-9444-4375-ae68-5b30e10f8c8a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.163024] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4680ef-4c2b-4d2b-9110-a7e92ba60d8c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.170690] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d342c4-59bd-4101-a875-9fad93fbc871 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.185398] env[61243]: DEBUG nova.compute.provider_tree [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1087.311296] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.311296] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.311296] env[61243]: DEBUG nova.network.neutron [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1087.326757] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Creating linked-clone VM from snapshot {{(pid=61243) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1087.327444] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-03d447aa-dca6-42db-b76a-0261d21d252a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.338438] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1087.338438] env[61243]: value = "task-1339518" [ 1087.338438] env[61243]: _type = "Task" [ 1087.338438] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.347849] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339518, 'name': CloneVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.460223] env[61243]: DEBUG nova.compute.manager [req-e7a6e4d6-b57a-436b-a494-39c487d6aa4b req-fb965aad-54b9-4904-845f-1363ef7cf3a2 service nova] [instance: fbf413f0-389c-4924-bf5d-400f95c89a84] Received event network-vif-deleted-b7573092-fe26-4633-ae34-f0b74fa0d3c7 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1087.689063] env[61243]: DEBUG nova.scheduler.client.report [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1087.850605] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339518, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.953898] env[61243]: DEBUG nova.compute.manager [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1087.980914] env[61243]: DEBUG nova.virt.hardware [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1087.981196] env[61243]: DEBUG nova.virt.hardware [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1087.981364] env[61243]: DEBUG nova.virt.hardware [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1087.981555] env[61243]: DEBUG nova.virt.hardware [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1087.981711] env[61243]: DEBUG nova.virt.hardware [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1087.981863] env[61243]: DEBUG nova.virt.hardware [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1087.982725] env[61243]: DEBUG nova.virt.hardware [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1087.983049] env[61243]: DEBUG nova.virt.hardware [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1087.983261] env[61243]: DEBUG nova.virt.hardware [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1087.983438] env[61243]: DEBUG nova.virt.hardware [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1087.983638] env[61243]: DEBUG nova.virt.hardware [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1087.984831] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b79f89-70d5-4363-be18-6484186abb9d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.993718] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5bac85e-0f05-439e-8cb4-1cf1b35a8733 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.007796] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Instance VIF info [] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1088.013180] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Creating folder: Project (eac00e596c23493f9d76f4b636b72256). Parent ref: group-v285636. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1088.013493] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cd8bdb28-05b7-45f0-bbbe-9e3edc06dca7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.025710] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Created folder: Project (eac00e596c23493f9d76f4b636b72256) in parent group-v285636. [ 1088.025919] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Creating folder: Instances. Parent ref: group-v285795. {{(pid=61243) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1088.026181] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-120d0ee1-6e3b-44ed-9d94-7c0c733cc37a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.036417] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Created folder: Instances in parent group-v285795. [ 1088.036654] env[61243]: DEBUG oslo.service.loopingcall [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1088.036873] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1088.037116] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0064d96a-4a35-4e8a-97e0-c86cddfe6ea4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.051335] env[61243]: DEBUG nova.network.neutron [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating instance_info_cache with network_info: [{"id": "f2a48425-d436-40c9-a2dc-694e6313da6d", "address": "fa:16:3e:34:84:e2", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2a48425-d4", "ovs_interfaceid": "f2a48425-d436-40c9-a2dc-694e6313da6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.055035] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1088.055035] env[61243]: value = "task-1339521" [ 1088.055035] env[61243]: _type = "Task" [ 1088.055035] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.062636] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339521, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.194552] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.263s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.195181] env[61243]: DEBUG nova.compute.manager [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1088.198751] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.948s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.198915] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.200826] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.462s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.222783] env[61243]: INFO nova.scheduler.client.report [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleted allocations for instance bb7c810f-91d1-43eb-b0cc-fb4c686b1f29 [ 1088.252908] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.253116] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.253299] env[61243]: DEBUG nova.network.neutron [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1088.349885] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339518, 'name': CloneVM_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.554278] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.568747] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339521, 'name': CreateVM_Task, 'duration_secs': 0.331586} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.568915] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1088.569330] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.569496] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.569828] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1088.570089] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abd6471a-308f-4795-aaf0-ef3ac7ef47b2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.574997] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1088.574997] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529ba528-ff9f-936f-940b-d5d6e6f0ef4c" [ 1088.574997] env[61243]: _type = "Task" [ 1088.574997] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.584211] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529ba528-ff9f-936f-940b-d5d6e6f0ef4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.704542] env[61243]: DEBUG nova.compute.utils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1088.711150] env[61243]: DEBUG nova.compute.manager [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1088.711325] env[61243]: DEBUG nova.network.neutron [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1088.731462] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2c76b027-20e9-4616-9241-9e67171088a6 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "bb7c810f-91d1-43eb-b0cc-fb4c686b1f29" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.590s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.759597] env[61243]: DEBUG nova.policy [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '785fd146024849efa0b67b1a7f731393', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9db958446334496bab832b95f2363f2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1088.850371] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339518, 'name': CloneVM_Task, 'duration_secs': 1.026721} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.850711] env[61243]: INFO nova.virt.vmwareapi.vmops [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Created linked-clone VM from snapshot [ 1088.851659] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fccc4b11-7959-4efd-8e27-00f035c4e303 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.859716] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Uploading image 713fcb79-3319-4eb2-bdaf-38d7d056d38b {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1088.881370] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1088.881370] env[61243]: value = "vm-285794" [ 1088.881370] env[61243]: _type = "VirtualMachine" [ 1088.881370] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1088.881630] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5e166ccf-fcc3-4dc8-bc05-6e3869f03f41 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.889445] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lease: (returnval){ [ 1088.889445] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5206f424-508f-a272-90a9-05a3d50402eb" [ 1088.889445] env[61243]: _type = "HttpNfcLease" [ 1088.889445] env[61243]: } obtained for exporting VM: (result){ [ 1088.889445] env[61243]: value = "vm-285794" [ 1088.889445] env[61243]: _type = "VirtualMachine" [ 1088.889445] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1088.889728] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the lease: (returnval){ [ 1088.889728] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5206f424-508f-a272-90a9-05a3d50402eb" [ 1088.889728] env[61243]: _type = "HttpNfcLease" [ 1088.889728] env[61243]: } to be ready. {{(pid=61243) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1088.897321] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1088.897321] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5206f424-508f-a272-90a9-05a3d50402eb" [ 1088.897321] env[61243]: _type = "HttpNfcLease" [ 1088.897321] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1088.926928] env[61243]: DEBUG nova.compute.manager [req-a0699af5-2b3e-4a20-8aef-c18f2c1f4060 req-d170611c-c238-4bf2-a563-671d006cc8cb service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Received event network-changed-de6bbdd5-c1ee-40d7-a91c-1938f53863a7 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1088.927124] env[61243]: DEBUG nova.compute.manager [req-a0699af5-2b3e-4a20-8aef-c18f2c1f4060 req-d170611c-c238-4bf2-a563-671d006cc8cb service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Refreshing instance network info cache due to event network-changed-de6bbdd5-c1ee-40d7-a91c-1938f53863a7. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1088.927326] env[61243]: DEBUG oslo_concurrency.lockutils [req-a0699af5-2b3e-4a20-8aef-c18f2c1f4060 req-d170611c-c238-4bf2-a563-671d006cc8cb service nova] Acquiring lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.065719] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d5f738e-c28a-4885-8165-d1281b6b1923 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.073573] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3865236d-3e42-4d03-896f-35aec2e77316 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.082527] env[61243]: INFO nova.network.neutron [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Port deadb902-8527-45ad-aa9c-bb805228869f from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1089.082864] env[61243]: DEBUG nova.network.neutron [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Updating instance_info_cache with network_info: [{"id": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "address": "fa:16:3e:43:9a:6f", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde6bbdd5-c1", "ovs_interfaceid": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.090300] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]529ba528-ff9f-936f-940b-d5d6e6f0ef4c, 'name': SearchDatastore_Task, 'duration_secs': 0.020845} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.095092] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.095342] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1089.095578] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.095736] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.095921] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1089.100418] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-80bb80b3-2528-42d0-89b5-7beef0d88f51 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.110236] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1089.110428] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1089.111162] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2122db09-8f65-4371-a73b-890bea66da0a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.118093] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1089.118093] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5229bba0-d2f1-40b8-ee58-2c10e25d5cb5" [ 1089.118093] env[61243]: _type = "Task" [ 1089.118093] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.126347] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5229bba0-d2f1-40b8-ee58-2c10e25d5cb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.167953] env[61243]: DEBUG nova.network.neutron [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Successfully created port: aa211a5b-6d6e-437d-9219-337bfb9be880 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1089.211838] env[61243]: DEBUG nova.compute.manager [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1089.215726] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Applying migration context for instance 60fa2863-f0aa-4798-8e5f-e93df47f93e8 as it has an incoming, in-progress migration c6475dac-275e-43dc-b28b-6d7f0683bc1a. Migration status is post-migrating {{(pid=61243) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1089.217045] env[61243]: INFO nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating resource usage from migration c6475dac-275e-43dc-b28b-6d7f0683bc1a [ 1089.241735] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 3f839aa5-2b9a-4807-b63b-931f74455532 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1089.241735] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 466f6a0a-9bc2-4709-8625-e6fc65b68f80 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1089.241735] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 486ebf65-4376-424c-9deb-f1ef70e44be3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1089.241735] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 8279496d-4a90-4962-9cc2-5d5e1ee33ff0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1089.241735] env[61243]: WARNING nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 26ca8850-4785-4c02-9dd6-3788d0164510 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1089.241735] env[61243]: WARNING nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance fbf413f0-389c-4924-bf5d-400f95c89a84 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1089.241735] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Migration c6475dac-275e-43dc-b28b-6d7f0683bc1a is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1089.241735] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 60fa2863-f0aa-4798-8e5f-e93df47f93e8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1089.241735] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 6786bf24-f760-4f9b-947a-25d19ae08253 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1089.241735] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 3d6acdec-ec3a-46b6-858a-135a734ea8ae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1089.401426] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1089.401426] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5206f424-508f-a272-90a9-05a3d50402eb" [ 1089.401426] env[61243]: _type = "HttpNfcLease" [ 1089.401426] env[61243]: } is ready. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1089.401426] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1089.401426] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5206f424-508f-a272-90a9-05a3d50402eb" [ 1089.401426] env[61243]: _type = "HttpNfcLease" [ 1089.401426] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1089.405022] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc63f00f-aded-4196-948b-79ef5f14cd41 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.413817] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c32add-bb86-2b55-fe68-48a2c777f698/disk-0.vmdk from lease info. {{(pid=61243) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1089.413817] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c32add-bb86-2b55-fe68-48a2c777f698/disk-0.vmdk for reading. {{(pid=61243) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1089.505948] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b901389f-a158-43b7-92f6-d1cbf2a8d4e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.585626] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.587829] env[61243]: DEBUG oslo_concurrency.lockutils [req-a0699af5-2b3e-4a20-8aef-c18f2c1f4060 req-d170611c-c238-4bf2-a563-671d006cc8cb service nova] Acquired lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.588061] env[61243]: DEBUG nova.network.neutron [req-a0699af5-2b3e-4a20-8aef-c18f2c1f4060 req-d170611c-c238-4bf2-a563-671d006cc8cb service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Refreshing network info cache for port de6bbdd5-c1ee-40d7-a91c-1938f53863a7 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1089.603299] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "interface-8279496d-4a90-4962-9cc2-5d5e1ee33ff0-deadb902-8527-45ad-aa9c-bb805228869f" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.603464] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-8279496d-4a90-4962-9cc2-5d5e1ee33ff0-deadb902-8527-45ad-aa9c-bb805228869f" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.603819] env[61243]: DEBUG nova.objects.instance [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lazy-loading 'flavor' on Instance uuid 8279496d-4a90-4962-9cc2-5d5e1ee33ff0 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.629874] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5229bba0-d2f1-40b8-ee58-2c10e25d5cb5, 'name': SearchDatastore_Task, 'duration_secs': 0.009383} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.630728] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffbc1707-9562-41df-a433-315753c702f1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.636316] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1089.636316] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52761f06-ab7c-36b3-9891-7b6735da78d0" [ 1089.636316] env[61243]: _type = "Task" [ 1089.636316] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.643836] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52761f06-ab7c-36b3-9891-7b6735da78d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.744369] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance d24f3683-169b-46a8-9387-b6fc88b654bd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1090.091161] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6c0b3aed-f14c-4156-9c83-a028f2aa2736 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-466f6a0a-9bc2-4709-8625-e6fc65b68f80-deadb902-8527-45ad-aa9c-bb805228869f" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.952s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.150646] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52761f06-ab7c-36b3-9891-7b6735da78d0, 'name': SearchDatastore_Task, 'duration_secs': 0.010274} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.151153] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.151644] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 6786bf24-f760-4f9b-947a-25d19ae08253/6786bf24-f760-4f9b-947a-25d19ae08253.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1090.152021] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bc16137b-767a-4582-9b24-54692996a355 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.164954] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1090.164954] env[61243]: value = "task-1339523" [ 1090.164954] env[61243]: _type = "Task" [ 1090.164954] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.173993] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339523, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.205445] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe6f9500-9d03-491f-aa4c-9d230be586de {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.231228] env[61243]: DEBUG nova.compute.manager [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1090.238495] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-727c03d5-d573-4672-9e69-f87e42137672 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.248824] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating instance '60fa2863-f0aa-4798-8e5f-e93df47f93e8' progress to 83 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1090.255645] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 91b14a22-811a-4151-9769-9c684ec993db has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1090.255974] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1090.256205] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1090.269411] env[61243]: DEBUG nova.virt.hardware [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1090.270405] env[61243]: DEBUG nova.virt.hardware [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1090.270621] env[61243]: DEBUG nova.virt.hardware [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1090.270829] env[61243]: DEBUG nova.virt.hardware [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1090.270990] env[61243]: DEBUG nova.virt.hardware [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1090.271171] env[61243]: DEBUG nova.virt.hardware [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1090.272324] env[61243]: DEBUG nova.virt.hardware [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1090.272543] env[61243]: DEBUG nova.virt.hardware [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1090.272739] env[61243]: DEBUG nova.virt.hardware [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1090.272918] env[61243]: DEBUG nova.virt.hardware [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1090.273131] env[61243]: DEBUG nova.virt.hardware [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1090.274563] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34447dec-7c4d-43cd-a08e-b6d36f75b7c3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.288926] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08df8005-c08c-4bda-ba73-37e6463ad1c6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.307084] env[61243]: DEBUG nova.objects.instance [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lazy-loading 'pci_requests' on Instance uuid 8279496d-4a90-4962-9cc2-5d5e1ee33ff0 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1090.482123] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c566d8-e96b-4fbc-8491-7edc44712d42 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.495066] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1488ec57-3f9f-4184-9478-8e037b59d145 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.500957] env[61243]: DEBUG nova.network.neutron [req-a0699af5-2b3e-4a20-8aef-c18f2c1f4060 req-d170611c-c238-4bf2-a563-671d006cc8cb service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Updated VIF entry in instance network info cache for port de6bbdd5-c1ee-40d7-a91c-1938f53863a7. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1090.502404] env[61243]: DEBUG nova.network.neutron [req-a0699af5-2b3e-4a20-8aef-c18f2c1f4060 req-d170611c-c238-4bf2-a563-671d006cc8cb service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Updating instance_info_cache with network_info: [{"id": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "address": "fa:16:3e:43:9a:6f", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde6bbdd5-c1", "ovs_interfaceid": "de6bbdd5-c1ee-40d7-a91c-1938f53863a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.537345] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f320edc4-a39f-4231-98c6-2c8a3e26d5a7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.549121] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab41b8b1-4ed9-4987-a37a-90de04f81182 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.574036] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1090.647938] env[61243]: DEBUG nova.compute.manager [req-c9ea022e-e89c-4f79-9293-46614f4cd9fe req-02287f61-5357-4d5f-97da-2f5fc3f6f9a9 service nova] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Received event network-vif-plugged-aa211a5b-6d6e-437d-9219-337bfb9be880 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1090.648202] env[61243]: DEBUG oslo_concurrency.lockutils [req-c9ea022e-e89c-4f79-9293-46614f4cd9fe req-02287f61-5357-4d5f-97da-2f5fc3f6f9a9 service nova] Acquiring lock "3d6acdec-ec3a-46b6-858a-135a734ea8ae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.648888] env[61243]: DEBUG oslo_concurrency.lockutils [req-c9ea022e-e89c-4f79-9293-46614f4cd9fe req-02287f61-5357-4d5f-97da-2f5fc3f6f9a9 service nova] Lock "3d6acdec-ec3a-46b6-858a-135a734ea8ae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.649106] env[61243]: DEBUG oslo_concurrency.lockutils [req-c9ea022e-e89c-4f79-9293-46614f4cd9fe req-02287f61-5357-4d5f-97da-2f5fc3f6f9a9 service nova] Lock "3d6acdec-ec3a-46b6-858a-135a734ea8ae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.649371] env[61243]: DEBUG nova.compute.manager [req-c9ea022e-e89c-4f79-9293-46614f4cd9fe req-02287f61-5357-4d5f-97da-2f5fc3f6f9a9 service nova] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] No waiting events found dispatching network-vif-plugged-aa211a5b-6d6e-437d-9219-337bfb9be880 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1090.649714] env[61243]: WARNING nova.compute.manager [req-c9ea022e-e89c-4f79-9293-46614f4cd9fe req-02287f61-5357-4d5f-97da-2f5fc3f6f9a9 service nova] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Received unexpected event network-vif-plugged-aa211a5b-6d6e-437d-9219-337bfb9be880 for instance with vm_state building and task_state spawning. [ 1090.679678] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339523, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.750627] env[61243]: DEBUG nova.network.neutron [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Successfully updated port: aa211a5b-6d6e-437d-9219-337bfb9be880 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1090.761197] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1090.761822] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-77d49b12-12b9-495a-96eb-f8ab1df309bd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.770699] env[61243]: DEBUG oslo_vmware.api [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1090.770699] env[61243]: value = "task-1339524" [ 1090.770699] env[61243]: _type = "Task" [ 1090.770699] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.779301] env[61243]: DEBUG oslo_vmware.api [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339524, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.815779] env[61243]: DEBUG nova.objects.base [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Object Instance<8279496d-4a90-4962-9cc2-5d5e1ee33ff0> lazy-loaded attributes: flavor,pci_requests {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1090.816054] env[61243]: DEBUG nova.network.neutron [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1090.888798] env[61243]: DEBUG nova.policy [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b32f4788825f4a8e82e028cc10347746', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '683c027e4adc4cc4a6b8747a8c8c0de9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1090.994670] env[61243]: DEBUG nova.compute.manager [req-beb070e8-ed30-4277-9ed9-239d29108da2 req-2b272d1e-6435-40a9-9a4a-4ef42b552ea1 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Received event network-changed-5d2d1a82-6eda-413f-9581-9fa9107be4bd {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1090.995238] env[61243]: DEBUG nova.compute.manager [req-beb070e8-ed30-4277-9ed9-239d29108da2 req-2b272d1e-6435-40a9-9a4a-4ef42b552ea1 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Refreshing instance network info cache due to event network-changed-5d2d1a82-6eda-413f-9581-9fa9107be4bd. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1090.995381] env[61243]: DEBUG oslo_concurrency.lockutils [req-beb070e8-ed30-4277-9ed9-239d29108da2 req-2b272d1e-6435-40a9-9a4a-4ef42b552ea1 service nova] Acquiring lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.995525] env[61243]: DEBUG oslo_concurrency.lockutils [req-beb070e8-ed30-4277-9ed9-239d29108da2 req-2b272d1e-6435-40a9-9a4a-4ef42b552ea1 service nova] Acquired lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.995858] env[61243]: DEBUG nova.network.neutron [req-beb070e8-ed30-4277-9ed9-239d29108da2 req-2b272d1e-6435-40a9-9a4a-4ef42b552ea1 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Refreshing network info cache for port 5d2d1a82-6eda-413f-9581-9fa9107be4bd {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1091.005560] env[61243]: DEBUG oslo_concurrency.lockutils [req-a0699af5-2b3e-4a20-8aef-c18f2c1f4060 req-d170611c-c238-4bf2-a563-671d006cc8cb service nova] Releasing lock "refresh_cache-466f6a0a-9bc2-4709-8625-e6fc65b68f80" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.077345] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1091.177643] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339523, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.543546} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.178329] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 6786bf24-f760-4f9b-947a-25d19ae08253/6786bf24-f760-4f9b-947a-25d19ae08253.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1091.178329] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1091.179775] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b64a10f-1f81-44ac-9d54-fd9f5a78fd04 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.188067] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1091.188067] env[61243]: value = "task-1339525" [ 1091.188067] env[61243]: _type = "Task" [ 1091.188067] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.196811] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339525, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.253740] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "refresh_cache-3d6acdec-ec3a-46b6-858a-135a734ea8ae" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.253740] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "refresh_cache-3d6acdec-ec3a-46b6-858a-135a734ea8ae" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.254824] env[61243]: DEBUG nova.network.neutron [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1091.282125] env[61243]: DEBUG oslo_vmware.api [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339524, 'name': PowerOnVM_Task, 'duration_secs': 0.452818} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.282613] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1091.282930] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8460b3e9-9315-4bd2-ae1a-556638893201 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating instance '60fa2863-f0aa-4798-8e5f-e93df47f93e8' progress to 100 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1091.582949] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61243) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1091.583287] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.382s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.583615] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.743s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.585501] env[61243]: INFO nova.compute.claims [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1091.588430] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.588607] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Cleaning up deleted instances {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1091.700435] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339525, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083238} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.700961] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1091.702182] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db892f71-af5a-4e17-9a86-73af0e06992b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.722744] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 6786bf24-f760-4f9b-947a-25d19ae08253/6786bf24-f760-4f9b-947a-25d19ae08253.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1091.727046] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-920478cb-cad2-49d4-800b-4344c02e2f20 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.748446] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1091.748446] env[61243]: value = "task-1339526" [ 1091.748446] env[61243]: _type = "Task" [ 1091.748446] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.762272] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339526, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.794354] env[61243]: DEBUG nova.network.neutron [req-beb070e8-ed30-4277-9ed9-239d29108da2 req-2b272d1e-6435-40a9-9a4a-4ef42b552ea1 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Updated VIF entry in instance network info cache for port 5d2d1a82-6eda-413f-9581-9fa9107be4bd. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1091.794899] env[61243]: DEBUG nova.network.neutron [req-beb070e8-ed30-4277-9ed9-239d29108da2 req-2b272d1e-6435-40a9-9a4a-4ef42b552ea1 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Updating instance_info_cache with network_info: [{"id": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "address": "fa:16:3e:a4:7d:cc", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d2d1a82-6e", "ovs_interfaceid": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.799575] env[61243]: DEBUG nova.network.neutron [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1091.939053] env[61243]: DEBUG nova.network.neutron [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Updating instance_info_cache with network_info: [{"id": "aa211a5b-6d6e-437d-9219-337bfb9be880", "address": "fa:16:3e:1b:a1:fe", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa211a5b-6d", "ovs_interfaceid": "aa211a5b-6d6e-437d-9219-337bfb9be880", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.105549] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] There are 45 instances to clean {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1092.105903] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 3fabef9e-13ff-41b6-a4eb-ee65b66c37fe] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.260192] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339526, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.286533] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c256f04-570a-48c7-bfc7-584562cad7f3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.294856] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da390ae4-32bd-42c1-8959-0c861a3afab1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.298630] env[61243]: DEBUG oslo_concurrency.lockutils [req-beb070e8-ed30-4277-9ed9-239d29108da2 req-2b272d1e-6435-40a9-9a4a-4ef42b552ea1 service nova] Releasing lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.327842] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b584fc74-0a6d-4798-82ed-e163521521f2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.336102] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a1c84b-938a-4cba-8fca-6a0267b167b7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.353668] env[61243]: DEBUG nova.compute.provider_tree [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1092.399680] env[61243]: DEBUG nova.network.neutron [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Successfully updated port: deadb902-8527-45ad-aa9c-bb805228869f {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1092.441414] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "refresh_cache-3d6acdec-ec3a-46b6-858a-135a734ea8ae" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.441827] env[61243]: DEBUG nova.compute.manager [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Instance network_info: |[{"id": "aa211a5b-6d6e-437d-9219-337bfb9be880", "address": "fa:16:3e:1b:a1:fe", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa211a5b-6d", "ovs_interfaceid": "aa211a5b-6d6e-437d-9219-337bfb9be880", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1092.442341] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1b:a1:fe', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a071ecf4-e713-4f97-9271-8c17952f6dee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aa211a5b-6d6e-437d-9219-337bfb9be880', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1092.450865] env[61243]: DEBUG oslo.service.loopingcall [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1092.451499] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1092.451744] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-527077e7-41f3-4b65-b894-d9f34eaa0c56 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.474542] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1092.474542] env[61243]: value = "task-1339527" [ 1092.474542] env[61243]: _type = "Task" [ 1092.474542] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.487408] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339527, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.612778] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: df141b12-1d34-4a47-b6bb-15cb75d266c3] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.687072] env[61243]: DEBUG nova.compute.manager [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Received event network-changed-aa211a5b-6d6e-437d-9219-337bfb9be880 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1092.687300] env[61243]: DEBUG nova.compute.manager [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Refreshing instance network info cache due to event network-changed-aa211a5b-6d6e-437d-9219-337bfb9be880. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1092.687517] env[61243]: DEBUG oslo_concurrency.lockutils [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] Acquiring lock "refresh_cache-3d6acdec-ec3a-46b6-858a-135a734ea8ae" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.687666] env[61243]: DEBUG oslo_concurrency.lockutils [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] Acquired lock "refresh_cache-3d6acdec-ec3a-46b6-858a-135a734ea8ae" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.687831] env[61243]: DEBUG nova.network.neutron [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Refreshing network info cache for port aa211a5b-6d6e-437d-9219-337bfb9be880 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1092.760237] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339526, 'name': ReconfigVM_Task, 'duration_secs': 0.570364} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.760561] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 6786bf24-f760-4f9b-947a-25d19ae08253/6786bf24-f760-4f9b-947a-25d19ae08253.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1092.761206] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65b182d2-3908-4731-9cf2-27d2099c7352 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.768793] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1092.768793] env[61243]: value = "task-1339528" [ 1092.768793] env[61243]: _type = "Task" [ 1092.768793] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.779575] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339528, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.861569] env[61243]: DEBUG nova.scheduler.client.report [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1092.902732] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.902925] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.903154] env[61243]: DEBUG nova.network.neutron [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1092.985924] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339527, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.116557] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: bb7c810f-91d1-43eb-b0cc-fb4c686b1f29] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.283045] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339528, 'name': Rename_Task, 'duration_secs': 0.409517} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.283459] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1093.283817] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aec43601-9200-44e5-a2f8-c5a4e7c50c76 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.293780] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1093.293780] env[61243]: value = "task-1339529" [ 1093.293780] env[61243]: _type = "Task" [ 1093.293780] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.302387] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339529, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.368275] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.785s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.369066] env[61243]: DEBUG nova.compute.manager [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1093.371927] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.812s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.373534] env[61243]: INFO nova.compute.claims [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1093.445271] env[61243]: DEBUG nova.network.neutron [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Updated VIF entry in instance network info cache for port aa211a5b-6d6e-437d-9219-337bfb9be880. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1093.445634] env[61243]: DEBUG nova.network.neutron [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Updating instance_info_cache with network_info: [{"id": "aa211a5b-6d6e-437d-9219-337bfb9be880", "address": "fa:16:3e:1b:a1:fe", "network": {"id": "b866f22e-7ad5-4d04-8c67-07dcba71edf2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1832904963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9db958446334496bab832b95f2363f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa211a5b-6d", "ovs_interfaceid": "aa211a5b-6d6e-437d-9219-337bfb9be880", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.455615] env[61243]: WARNING nova.network.neutron [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] 7724b87f-3721-42a5-81ce-27ae98fc99c2 already exists in list: networks containing: ['7724b87f-3721-42a5-81ce-27ae98fc99c2']. ignoring it [ 1093.486328] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339527, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.623028] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 05d90648-c6ac-4ee9-ad4e-a3d2bf0be20a] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.721907] env[61243]: DEBUG nova.network.neutron [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Port f2a48425-d436-40c9-a2dc-694e6313da6d binding to destination host cpu-1 is already ACTIVE {{(pid=61243) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1093.722305] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.722583] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.722846] env[61243]: DEBUG nova.network.neutron [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1093.805275] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339529, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.828161] env[61243]: DEBUG nova.network.neutron [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Updating instance_info_cache with network_info: [{"id": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "address": "fa:16:3e:a4:7d:cc", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d2d1a82-6e", "ovs_interfaceid": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "deadb902-8527-45ad-aa9c-bb805228869f", "address": "fa:16:3e:17:a0:99", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdeadb902-85", "ovs_interfaceid": "deadb902-8527-45ad-aa9c-bb805228869f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.877881] env[61243]: DEBUG nova.compute.utils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1093.881514] env[61243]: DEBUG nova.compute.manager [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1093.881685] env[61243]: DEBUG nova.network.neutron [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1093.922131] env[61243]: DEBUG nova.policy [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32d9e3276f974c8f87573270909076ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7651b565d1647eb948fe07faba32a41', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1093.947944] env[61243]: DEBUG oslo_concurrency.lockutils [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] Releasing lock "refresh_cache-3d6acdec-ec3a-46b6-858a-135a734ea8ae" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.948190] env[61243]: DEBUG nova.compute.manager [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Received event network-vif-plugged-deadb902-8527-45ad-aa9c-bb805228869f {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1093.948398] env[61243]: DEBUG oslo_concurrency.lockutils [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] Acquiring lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.948604] env[61243]: DEBUG oslo_concurrency.lockutils [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] Lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.948774] env[61243]: DEBUG oslo_concurrency.lockutils [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] Lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.948950] env[61243]: DEBUG nova.compute.manager [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] No waiting events found dispatching network-vif-plugged-deadb902-8527-45ad-aa9c-bb805228869f {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1093.949132] env[61243]: WARNING nova.compute.manager [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Received unexpected event network-vif-plugged-deadb902-8527-45ad-aa9c-bb805228869f for instance with vm_state active and task_state None. [ 1093.949304] env[61243]: DEBUG nova.compute.manager [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Received event network-changed-deadb902-8527-45ad-aa9c-bb805228869f {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1093.949466] env[61243]: DEBUG nova.compute.manager [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Refreshing instance network info cache due to event network-changed-deadb902-8527-45ad-aa9c-bb805228869f. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1093.949638] env[61243]: DEBUG oslo_concurrency.lockutils [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] Acquiring lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.988028] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339527, 'name': CreateVM_Task, 'duration_secs': 1.04867} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.988230] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1093.988972] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.989168] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.989520] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1093.989793] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a197e199-5f13-4a14-a566-80b328e1a3fd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.996153] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1093.996153] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5243cc29-6e65-3407-2745-bfce66f6ae15" [ 1093.996153] env[61243]: _type = "Task" [ 1093.996153] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.005321] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5243cc29-6e65-3407-2745-bfce66f6ae15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.124196] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 0e03eee3-e7bc-483f-948f-a5b408375e24] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.220970] env[61243]: DEBUG nova.network.neutron [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Successfully created port: 1c517f1d-301b-43a0-8ae9-6099f75580ed {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1094.306904] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339529, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.331769] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.332533] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.332712] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.333012] env[61243]: DEBUG oslo_concurrency.lockutils [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] Acquired lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.333234] env[61243]: DEBUG nova.network.neutron [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Refreshing network info cache for port deadb902-8527-45ad-aa9c-bb805228869f {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1094.335139] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a28c8e-2a8f-496f-9c52-b60c18c84eee {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.358815] env[61243]: DEBUG nova.virt.hardware [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1094.359207] env[61243]: DEBUG nova.virt.hardware [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1094.359434] env[61243]: DEBUG nova.virt.hardware [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1094.359689] env[61243]: DEBUG nova.virt.hardware [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1094.359902] env[61243]: DEBUG nova.virt.hardware [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1094.360127] env[61243]: DEBUG nova.virt.hardware [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1094.360457] env[61243]: DEBUG nova.virt.hardware [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1094.360703] env[61243]: DEBUG nova.virt.hardware [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1094.360941] env[61243]: DEBUG nova.virt.hardware [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1094.361182] env[61243]: DEBUG nova.virt.hardware [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1094.361450] env[61243]: DEBUG nova.virt.hardware [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1094.368420] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Reconfiguring VM to attach interface {{(pid=61243) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1094.369214] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9ba30ed-47ea-4c13-8b01-3b3610728436 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.381746] env[61243]: DEBUG nova.compute.manager [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1094.395037] env[61243]: DEBUG oslo_vmware.api [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1094.395037] env[61243]: value = "task-1339530" [ 1094.395037] env[61243]: _type = "Task" [ 1094.395037] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.408532] env[61243]: DEBUG oslo_vmware.api [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339530, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.511476] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5243cc29-6e65-3407-2745-bfce66f6ae15, 'name': SearchDatastore_Task, 'duration_secs': 0.191747} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.512688] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.513242] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1094.513503] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.513808] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.514145] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1094.514485] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7ce4b56f-d427-4d6c-b0a8-31fab89b96a1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.621193] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faec5753-5443-41db-9b6f-b981a81be59e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.627987] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: dca3301d-ef3e-48ef-920c-866b4a086ea6] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.631244] env[61243]: DEBUG nova.network.neutron [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating instance_info_cache with network_info: [{"id": "f2a48425-d436-40c9-a2dc-694e6313da6d", "address": "fa:16:3e:34:84:e2", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2a48425-d4", "ovs_interfaceid": "f2a48425-d436-40c9-a2dc-694e6313da6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.635213] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90a5454-acc7-4d97-9d9c-2351144ca71e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.668140] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c8ec6a-c8ff-40fc-997d-5683a2d89198 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.677914] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31243ac9-8336-4da6-85f9-b33a2ac8f0c3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.694838] env[61243]: DEBUG nova.compute.provider_tree [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1094.771357] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1094.771561] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1094.772327] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37bc6272-9d0a-41ac-ab9f-b3c6388217a2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.779017] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1094.779017] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c5637b-fba2-bcc8-f695-47fc50a39c7d" [ 1094.779017] env[61243]: _type = "Task" [ 1094.779017] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.787562] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c5637b-fba2-bcc8-f695-47fc50a39c7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.805462] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339529, 'name': PowerOnVM_Task} progress is 70%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.907162] env[61243]: DEBUG oslo_vmware.api [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339530, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.055578] env[61243]: DEBUG nova.network.neutron [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Updated VIF entry in instance network info cache for port deadb902-8527-45ad-aa9c-bb805228869f. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1095.056017] env[61243]: DEBUG nova.network.neutron [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Updating instance_info_cache with network_info: [{"id": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "address": "fa:16:3e:a4:7d:cc", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d2d1a82-6e", "ovs_interfaceid": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "deadb902-8527-45ad-aa9c-bb805228869f", "address": "fa:16:3e:17:a0:99", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdeadb902-85", "ovs_interfaceid": "deadb902-8527-45ad-aa9c-bb805228869f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.136237] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 2c4217cd-af4a-4d64-a422-73955326df03] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.139156] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.198879] env[61243]: DEBUG nova.scheduler.client.report [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1095.292260] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52c5637b-fba2-bcc8-f695-47fc50a39c7d, 'name': SearchDatastore_Task, 'duration_secs': 0.022696} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.293439] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27095480-c5b5-439b-aa43-bc61618694f2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.304831] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1095.304831] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52640537-8934-726e-33e2-f9aef39be204" [ 1095.304831] env[61243]: _type = "Task" [ 1095.304831] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.308199] env[61243]: DEBUG oslo_vmware.api [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339529, 'name': PowerOnVM_Task, 'duration_secs': 2.007054} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.311224] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1095.311453] env[61243]: INFO nova.compute.manager [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Took 7.36 seconds to spawn the instance on the hypervisor. [ 1095.311710] env[61243]: DEBUG nova.compute.manager [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1095.312406] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a1645b8-eb1a-4d26-8f7a-c77695258901 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.324849] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52640537-8934-726e-33e2-f9aef39be204, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.395249] env[61243]: DEBUG nova.compute.manager [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1095.408612] env[61243]: DEBUG oslo_vmware.api [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339530, 'name': ReconfigVM_Task, 'duration_secs': 0.683413} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.409192] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.409432] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Reconfigured VM to attach interface {{(pid=61243) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1095.422285] env[61243]: DEBUG nova.virt.hardware [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1095.422585] env[61243]: DEBUG nova.virt.hardware [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1095.422755] env[61243]: DEBUG nova.virt.hardware [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1095.422942] env[61243]: DEBUG nova.virt.hardware [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1095.423119] env[61243]: DEBUG nova.virt.hardware [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1095.423280] env[61243]: DEBUG nova.virt.hardware [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1095.423570] env[61243]: DEBUG nova.virt.hardware [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1095.423720] env[61243]: DEBUG nova.virt.hardware [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1095.423898] env[61243]: DEBUG nova.virt.hardware [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1095.424080] env[61243]: DEBUG nova.virt.hardware [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1095.424262] env[61243]: DEBUG nova.virt.hardware [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1095.425193] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f395d19-caf6-437b-b448-014c6ba93327 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.434415] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ec414f-83be-4284-8b61-2775103483e7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.558878] env[61243]: DEBUG oslo_concurrency.lockutils [req-b9cfe3f1-88a1-47fc-b086-f9cd707474d2 req-438e6214-c807-4ee9-99f9-6a2cc6a4909e service nova] Releasing lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.639708] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 5f1c246d-2732-4da4-a00b-b92e2990aa67] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.643344] env[61243]: DEBUG nova.compute.manager [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61243) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1095.703351] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.703922] env[61243]: DEBUG nova.compute.manager [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1095.710017] env[61243]: DEBUG oslo_concurrency.lockutils [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.787s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.710017] env[61243]: DEBUG oslo_concurrency.lockutils [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.711508] env[61243]: DEBUG oslo_concurrency.lockutils [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.763s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.711702] env[61243]: DEBUG oslo_concurrency.lockutils [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.737585] env[61243]: INFO nova.scheduler.client.report [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Deleted allocations for instance 26ca8850-4785-4c02-9dd6-3788d0164510 [ 1095.740357] env[61243]: INFO nova.scheduler.client.report [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Deleted allocations for instance fbf413f0-389c-4924-bf5d-400f95c89a84 [ 1095.819638] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52640537-8934-726e-33e2-f9aef39be204, 'name': SearchDatastore_Task, 'duration_secs': 0.017853} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.820603] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.821065] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 3d6acdec-ec3a-46b6-858a-135a734ea8ae/3d6acdec-ec3a-46b6-858a-135a734ea8ae.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1095.821499] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea5c9644-0b8d-4769-add1-362c2ec95022 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.830626] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1095.830626] env[61243]: value = "task-1339531" [ 1095.830626] env[61243]: _type = "Task" [ 1095.830626] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.839145] env[61243]: INFO nova.compute.manager [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Took 18.45 seconds to build instance. [ 1095.844758] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339531, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.914190] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2ff02cc9-e045-4152-ba24-48ac64f161c1 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-8279496d-4a90-4962-9cc2-5d5e1ee33ff0-deadb902-8527-45ad-aa9c-bb805228869f" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.310s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.917880] env[61243]: DEBUG nova.compute.manager [req-aed883d0-6f64-4e42-b642-918f5b47215c req-3b996e8f-744a-44e9-9a5b-f04a60c6bae4 service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Received event network-vif-plugged-1c517f1d-301b-43a0-8ae9-6099f75580ed {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1095.918158] env[61243]: DEBUG oslo_concurrency.lockutils [req-aed883d0-6f64-4e42-b642-918f5b47215c req-3b996e8f-744a-44e9-9a5b-f04a60c6bae4 service nova] Acquiring lock "d24f3683-169b-46a8-9387-b6fc88b654bd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.918411] env[61243]: DEBUG oslo_concurrency.lockutils [req-aed883d0-6f64-4e42-b642-918f5b47215c req-3b996e8f-744a-44e9-9a5b-f04a60c6bae4 service nova] Lock "d24f3683-169b-46a8-9387-b6fc88b654bd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.918596] env[61243]: DEBUG oslo_concurrency.lockutils [req-aed883d0-6f64-4e42-b642-918f5b47215c req-3b996e8f-744a-44e9-9a5b-f04a60c6bae4 service nova] Lock "d24f3683-169b-46a8-9387-b6fc88b654bd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.918771] env[61243]: DEBUG nova.compute.manager [req-aed883d0-6f64-4e42-b642-918f5b47215c req-3b996e8f-744a-44e9-9a5b-f04a60c6bae4 service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] No waiting events found dispatching network-vif-plugged-1c517f1d-301b-43a0-8ae9-6099f75580ed {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1095.918968] env[61243]: WARNING nova.compute.manager [req-aed883d0-6f64-4e42-b642-918f5b47215c req-3b996e8f-744a-44e9-9a5b-f04a60c6bae4 service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Received unexpected event network-vif-plugged-1c517f1d-301b-43a0-8ae9-6099f75580ed for instance with vm_state building and task_state spawning. [ 1095.959404] env[61243]: DEBUG nova.network.neutron [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Successfully updated port: 1c517f1d-301b-43a0-8ae9-6099f75580ed {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1096.147112] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 818476f0-e928-44d5-b3d4-101b7a05c9aa] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.216164] env[61243]: DEBUG nova.compute.utils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1096.217826] env[61243]: DEBUG nova.compute.manager [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1096.218157] env[61243]: DEBUG nova.network.neutron [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1096.253502] env[61243]: DEBUG oslo_concurrency.lockutils [None req-736ae820-97f9-4d56-9c39-95b33d44a2c4 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "26ca8850-4785-4c02-9dd6-3788d0164510" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.371s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.254911] env[61243]: DEBUG oslo_concurrency.lockutils [None req-083e1c35-688d-407a-a5fb-ed47e5e06c94 tempest-ListServersNegativeTestJSON-676496071 tempest-ListServersNegativeTestJSON-676496071-project-member] Lock "fbf413f0-389c-4924-bf5d-400f95c89a84" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.309s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.308194] env[61243]: DEBUG nova.policy [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b9748805e2c48e4a8fd0354e916aa0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '324ed556b6f5419fa0376a7cfc508827', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1096.350964] env[61243]: DEBUG oslo_concurrency.lockutils [None req-4483ce86-451d-4a94-8cb2-3fd27967adc2 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Lock "6786bf24-f760-4f9b-947a-25d19ae08253" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.968s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.350964] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339531, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.464988] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "refresh_cache-d24f3683-169b-46a8-9387-b6fc88b654bd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.464988] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquired lock "refresh_cache-d24f3683-169b-46a8-9387-b6fc88b654bd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.464988] env[61243]: DEBUG nova.network.neutron [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1096.639449] env[61243]: INFO nova.compute.manager [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Rebuilding instance [ 1096.650627] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 5660de64-7152-4832-960e-2deb1c0bc37e] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.699054] env[61243]: DEBUG nova.compute.manager [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1096.699054] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f911796f-b6c0-43de-9a82-7a3fd4044b33 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.724682] env[61243]: DEBUG nova.compute.manager [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1096.782186] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.782186] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.802926] env[61243]: DEBUG nova.network.neutron [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Successfully created port: 41768d32-20ce-414e-9c85-be16e03d795a {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1096.847937] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339531, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.706456} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.848341] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore1] 3d6acdec-ec3a-46b6-858a-135a734ea8ae/3d6acdec-ec3a-46b6-858a-135a734ea8ae.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1096.848650] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1096.848984] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2dbb17c2-bf10-491f-bfe0-0228134468c3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.861878] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1096.861878] env[61243]: value = "task-1339532" [ 1096.861878] env[61243]: _type = "Task" [ 1096.861878] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.883740] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339532, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.018760] env[61243]: DEBUG nova.network.neutron [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1097.154410] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 55f24903-b0b4-4124-afde-f31b2680a53a] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.211156] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1097.211534] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-21f3c21d-4f7f-49b6-bffe-73f3048745d3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.219335] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1097.219335] env[61243]: value = "task-1339533" [ 1097.219335] env[61243]: _type = "Task" [ 1097.219335] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.232998] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339533, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.287900] env[61243]: DEBUG nova.objects.instance [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lazy-loading 'migration_context' on Instance uuid 60fa2863-f0aa-4798-8e5f-e93df47f93e8 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.375230] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "interface-8279496d-4a90-4962-9cc2-5d5e1ee33ff0-deadb902-8527-45ad-aa9c-bb805228869f" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.376109] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-8279496d-4a90-4962-9cc2-5d5e1ee33ff0-deadb902-8527-45ad-aa9c-bb805228869f" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.377521] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339532, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.530025] env[61243]: DEBUG nova.network.neutron [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Updating instance_info_cache with network_info: [{"id": "1c517f1d-301b-43a0-8ae9-6099f75580ed", "address": "fa:16:3e:aa:38:1f", "network": {"id": "523d4cf1-467c-4ea6-80a4-ebbd01016efa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-919325376-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7651b565d1647eb948fe07faba32a41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c517f1d-30", "ovs_interfaceid": "1c517f1d-301b-43a0-8ae9-6099f75580ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.581045] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c32add-bb86-2b55-fe68-48a2c777f698/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1097.581045] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd709ffa-5148-4899-b905-0c32e76fd877 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.590161] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c32add-bb86-2b55-fe68-48a2c777f698/disk-0.vmdk is in state: ready. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1097.590161] env[61243]: ERROR oslo_vmware.rw_handles [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c32add-bb86-2b55-fe68-48a2c777f698/disk-0.vmdk due to incomplete transfer. [ 1097.590161] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1d17c6ca-dc7e-47b5-9739-93f07802846d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.596509] env[61243]: DEBUG oslo_vmware.rw_handles [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c32add-bb86-2b55-fe68-48a2c777f698/disk-0.vmdk. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1097.596721] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Uploaded image 713fcb79-3319-4eb2-bdaf-38d7d056d38b to the Glance image server {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1097.599225] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Destroying the VM {{(pid=61243) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1097.599446] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-33605bd4-2188-4273-9ecc-d9f83fc2b9c7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.608873] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1097.608873] env[61243]: value = "task-1339534" [ 1097.608873] env[61243]: _type = "Task" [ 1097.608873] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.617249] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339534, 'name': Destroy_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.659040] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 5103569e-a281-4e89-af69-3a070ab17564] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.729495] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339533, 'name': PowerOffVM_Task, 'duration_secs': 0.214305} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.729790] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1097.730027] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1097.730795] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce807bbd-4e1f-4782-a198-c40a92ff8943 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.734649] env[61243]: DEBUG nova.compute.manager [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1097.739016] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1097.739298] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-406438bc-605d-4a25-888a-048e22878ee1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.769782] env[61243]: DEBUG nova.virt.hardware [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1097.770044] env[61243]: DEBUG nova.virt.hardware [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1097.772089] env[61243]: DEBUG nova.virt.hardware [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1097.772089] env[61243]: DEBUG nova.virt.hardware [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1097.772089] env[61243]: DEBUG nova.virt.hardware [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1097.772089] env[61243]: DEBUG nova.virt.hardware [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1097.772089] env[61243]: DEBUG nova.virt.hardware [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1097.772089] env[61243]: DEBUG nova.virt.hardware [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1097.772089] env[61243]: DEBUG nova.virt.hardware [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1097.772089] env[61243]: DEBUG nova.virt.hardware [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1097.772089] env[61243]: DEBUG nova.virt.hardware [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1097.772706] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-284ccf50-1438-4a5d-879b-95c328426119 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.776603] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1097.776810] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1097.776992] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Deleting the datastore file [datastore1] 6786bf24-f760-4f9b-947a-25d19ae08253 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1097.780036] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0a114e54-ab6f-4610-b185-b909f54b9d77 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.784947] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29acedf7-5fdd-47d8-9298-6aab28febf95 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.788028] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1097.788028] env[61243]: value = "task-1339536" [ 1097.788028] env[61243]: _type = "Task" [ 1097.788028] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.807886] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339536, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.872563] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339532, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.912558} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.872907] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1097.873787] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d902d0f-bae9-46f4-8476-c8a6f6c9a4f3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.891801] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.891997] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.900645] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 3d6acdec-ec3a-46b6-858a-135a734ea8ae/3d6acdec-ec3a-46b6-858a-135a734ea8ae.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1097.905017] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1abd7fdf-af2b-4c6b-b25c-5ad2878848c4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.908225] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06470833-3beb-4041-ab13-daa87800cba4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.941458] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03dbce17-4043-48c8-ac01-b8d3a5b25041 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.944159] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1097.944159] env[61243]: value = "task-1339537" [ 1097.944159] env[61243]: _type = "Task" [ 1097.944159] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.952600] env[61243]: DEBUG nova.compute.manager [req-d7a62b28-25bb-4011-bc2d-8580d1b5a18b req-3f9e203b-16d1-4384-807f-482151c8aae1 service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Received event network-changed-1c517f1d-301b-43a0-8ae9-6099f75580ed {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1097.952857] env[61243]: DEBUG nova.compute.manager [req-d7a62b28-25bb-4011-bc2d-8580d1b5a18b req-3f9e203b-16d1-4384-807f-482151c8aae1 service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Refreshing instance network info cache due to event network-changed-1c517f1d-301b-43a0-8ae9-6099f75580ed. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1097.953134] env[61243]: DEBUG oslo_concurrency.lockutils [req-d7a62b28-25bb-4011-bc2d-8580d1b5a18b req-3f9e203b-16d1-4384-807f-482151c8aae1 service nova] Acquiring lock "refresh_cache-d24f3683-169b-46a8-9387-b6fc88b654bd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.975477] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Reconfiguring VM to detach interface {{(pid=61243) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1097.980758] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e16cbe5a-d1e2-4d93-a2aa-5b1587056440 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.996319] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339537, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.003291] env[61243]: DEBUG oslo_vmware.api [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1098.003291] env[61243]: value = "task-1339538" [ 1098.003291] env[61243]: _type = "Task" [ 1098.003291] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.013059] env[61243]: DEBUG oslo_vmware.api [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339538, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.033049] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Releasing lock "refresh_cache-d24f3683-169b-46a8-9387-b6fc88b654bd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.033426] env[61243]: DEBUG nova.compute.manager [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Instance network_info: |[{"id": "1c517f1d-301b-43a0-8ae9-6099f75580ed", "address": "fa:16:3e:aa:38:1f", "network": {"id": "523d4cf1-467c-4ea6-80a4-ebbd01016efa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-919325376-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7651b565d1647eb948fe07faba32a41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c517f1d-30", "ovs_interfaceid": "1c517f1d-301b-43a0-8ae9-6099f75580ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1098.033763] env[61243]: DEBUG oslo_concurrency.lockutils [req-d7a62b28-25bb-4011-bc2d-8580d1b5a18b req-3f9e203b-16d1-4384-807f-482151c8aae1 service nova] Acquired lock "refresh_cache-d24f3683-169b-46a8-9387-b6fc88b654bd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.033952] env[61243]: DEBUG nova.network.neutron [req-d7a62b28-25bb-4011-bc2d-8580d1b5a18b req-3f9e203b-16d1-4384-807f-482151c8aae1 service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Refreshing network info cache for port 1c517f1d-301b-43a0-8ae9-6099f75580ed {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1098.035661] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:38:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '42f08482-a1da-405d-9918-d733d9f5173c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1c517f1d-301b-43a0-8ae9-6099f75580ed', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1098.043401] env[61243]: DEBUG oslo.service.loopingcall [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1098.044573] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1098.045344] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d909d3b-bc4f-4cbc-aee1-dacc7bfe9894 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.048874] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-681bcf91-f0ae-41ba-810f-5006b464803b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.070524] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f3e6ad-a8ef-4e28-9377-95f21576e928 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.075080] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1098.075080] env[61243]: value = "task-1339539" [ 1098.075080] env[61243]: _type = "Task" [ 1098.075080] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.105294] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14208c5c-ef63-43f0-a8aa-9a21ad05ee28 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.112208] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339539, 'name': CreateVM_Task} progress is 15%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.122913] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8e545a-9651-4f82-b332-a8e3c97c50cc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.855647] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 7ca74bdc-c272-4b6e-a4dd-56d97144b197] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.859944] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339534, 'name': Destroy_Task, 'duration_secs': 0.456364} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.860820] env[61243]: DEBUG nova.network.neutron [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Successfully updated port: 41768d32-20ce-414e-9c85-be16e03d795a {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1098.873274] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Destroyed the VM [ 1098.873643] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Deleting Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1098.882297] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c41da0f8-03a2-4ea9-af5e-157be4e235e6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.885290] env[61243]: DEBUG nova.compute.provider_tree [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.902744] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339539, 'name': CreateVM_Task, 'duration_secs': 0.3863} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.903045] env[61243]: DEBUG oslo_vmware.api [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.903313] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339537, 'name': ReconfigVM_Task, 'duration_secs': 0.340126} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.903517] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339536, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.331347} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.904667] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1098.904960] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 3d6acdec-ec3a-46b6-858a-135a734ea8ae/3d6acdec-ec3a-46b6-858a-135a734ea8ae.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1098.905927] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1098.906151] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1098.906344] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1098.908738] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1098.908738] env[61243]: value = "task-1339540" [ 1098.908738] env[61243]: _type = "Task" [ 1098.908738] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.909331] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.909500] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.909822] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1098.910074] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-00d332b5-d600-417c-8abf-bc308ebf31b0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.914380] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe1b4403-61fc-425f-aa8d-bfe5bc7761b0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.923993] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1098.923993] env[61243]: value = "task-1339541" [ 1098.923993] env[61243]: _type = "Task" [ 1098.923993] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.927796] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339540, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.928411] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1098.928411] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ac9e51-670c-638c-ce87-4d21efa9c3ac" [ 1098.928411] env[61243]: _type = "Task" [ 1098.928411] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.940595] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339541, 'name': Rename_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.944397] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ac9e51-670c-638c-ce87-4d21efa9c3ac, 'name': SearchDatastore_Task} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.945138] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.945417] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1098.945690] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.945862] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.946111] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1098.946357] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d3cbd97f-29e6-498e-9825-731d1a78ab1b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.956136] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1098.956335] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1098.957096] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd958737-927b-4891-893d-f4c201d7078a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.966248] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1098.966248] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52412af4-32cb-a366-5e79-d42d4d45e99f" [ 1098.966248] env[61243]: _type = "Task" [ 1098.966248] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.975604] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52412af4-32cb-a366-5e79-d42d4d45e99f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.108779] env[61243]: DEBUG nova.network.neutron [req-d7a62b28-25bb-4011-bc2d-8580d1b5a18b req-3f9e203b-16d1-4384-807f-482151c8aae1 service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Updated VIF entry in instance network info cache for port 1c517f1d-301b-43a0-8ae9-6099f75580ed. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1099.109322] env[61243]: DEBUG nova.network.neutron [req-d7a62b28-25bb-4011-bc2d-8580d1b5a18b req-3f9e203b-16d1-4384-807f-482151c8aae1 service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Updating instance_info_cache with network_info: [{"id": "1c517f1d-301b-43a0-8ae9-6099f75580ed", "address": "fa:16:3e:aa:38:1f", "network": {"id": "523d4cf1-467c-4ea6-80a4-ebbd01016efa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-919325376-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7651b565d1647eb948fe07faba32a41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c517f1d-30", "ovs_interfaceid": "1c517f1d-301b-43a0-8ae9-6099f75580ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.363025] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: db6f3da1-7dc3-4b5c-a3af-d9d1c1774a09] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.370620] env[61243]: DEBUG oslo_vmware.api [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.375306] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "refresh_cache-91b14a22-811a-4151-9769-9c684ec993db" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1099.375919] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "refresh_cache-91b14a22-811a-4151-9769-9c684ec993db" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.375919] env[61243]: DEBUG nova.network.neutron [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1099.390081] env[61243]: DEBUG nova.scheduler.client.report [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1099.426357] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339540, 'name': RemoveSnapshot_Task, 'duration_secs': 0.403484} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.426633] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Deleted Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1099.426929] env[61243]: DEBUG nova.compute.manager [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1099.427978] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a104e330-9abb-41f6-9669-c5cf3356dcf1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.442022] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339541, 'name': Rename_Task, 'duration_secs': 0.156041} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.442431] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1099.442560] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3e66986-612e-4e08-8347-29a920e61f1a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.450235] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1099.450235] env[61243]: value = "task-1339542" [ 1099.450235] env[61243]: _type = "Task" [ 1099.450235] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.458844] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339542, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.476655] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52412af4-32cb-a366-5e79-d42d4d45e99f, 'name': SearchDatastore_Task, 'duration_secs': 0.011058} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.477459] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01f3350e-d1b0-40ba-853c-b3e68e2e2881 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.483506] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1099.483506] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52af3d8d-015c-fecb-2cb5-6b5237b3f1bf" [ 1099.483506] env[61243]: _type = "Task" [ 1099.483506] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.491689] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52af3d8d-015c-fecb-2cb5-6b5237b3f1bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.612801] env[61243]: DEBUG oslo_concurrency.lockutils [req-d7a62b28-25bb-4011-bc2d-8580d1b5a18b req-3f9e203b-16d1-4384-807f-482151c8aae1 service nova] Releasing lock "refresh_cache-d24f3683-169b-46a8-9387-b6fc88b654bd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1099.865718] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 7ce043e6-abdf-4708-80aa-221870375c23] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.871768] env[61243]: DEBUG oslo_vmware.api [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.908962] env[61243]: DEBUG nova.network.neutron [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1099.944972] env[61243]: INFO nova.compute.manager [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Shelve offloading [ 1099.946629] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1099.946879] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9108543a-2eef-4d63-8241-c88fda8c753e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.950654] env[61243]: DEBUG nova.virt.hardware [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1099.950822] env[61243]: DEBUG nova.virt.hardware [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1099.950987] env[61243]: DEBUG nova.virt.hardware [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1099.951196] env[61243]: DEBUG nova.virt.hardware [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1099.951348] env[61243]: DEBUG nova.virt.hardware [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1099.951626] env[61243]: DEBUG nova.virt.hardware [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1099.951856] env[61243]: DEBUG nova.virt.hardware [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1099.952067] env[61243]: DEBUG nova.virt.hardware [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1099.952204] env[61243]: DEBUG nova.virt.hardware [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1099.953043] env[61243]: DEBUG nova.virt.hardware [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1099.953043] env[61243]: DEBUG nova.virt.hardware [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1099.953514] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365090f9-84f9-49b3-9c46-b37fa72c0c22 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.963497] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1099.963497] env[61243]: value = "task-1339543" [ 1099.963497] env[61243]: _type = "Task" [ 1099.963497] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.968116] env[61243]: DEBUG oslo_vmware.api [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339542, 'name': PowerOnVM_Task, 'duration_secs': 0.471539} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.971404] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1099.971701] env[61243]: INFO nova.compute.manager [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Took 9.74 seconds to spawn the instance on the hypervisor. [ 1099.971860] env[61243]: DEBUG nova.compute.manager [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1099.973075] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8327c9c9-bd06-4f32-9a32-14d8e12a105d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.978281] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea68aaf-7eba-4d6a-a029-d175a745dceb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.981524] env[61243]: DEBUG nova.compute.manager [req-e050b23a-640e-4f9d-8f2f-b23fdf12a4c4 req-9f2295bd-b6f3-49dc-b3ef-f4ab35d949d2 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Received event network-vif-plugged-41768d32-20ce-414e-9c85-be16e03d795a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1099.981738] env[61243]: DEBUG oslo_concurrency.lockutils [req-e050b23a-640e-4f9d-8f2f-b23fdf12a4c4 req-9f2295bd-b6f3-49dc-b3ef-f4ab35d949d2 service nova] Acquiring lock "91b14a22-811a-4151-9769-9c684ec993db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.981951] env[61243]: DEBUG oslo_concurrency.lockutils [req-e050b23a-640e-4f9d-8f2f-b23fdf12a4c4 req-9f2295bd-b6f3-49dc-b3ef-f4ab35d949d2 service nova] Lock "91b14a22-811a-4151-9769-9c684ec993db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.982149] env[61243]: DEBUG oslo_concurrency.lockutils [req-e050b23a-640e-4f9d-8f2f-b23fdf12a4c4 req-9f2295bd-b6f3-49dc-b3ef-f4ab35d949d2 service nova] Lock "91b14a22-811a-4151-9769-9c684ec993db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.982382] env[61243]: DEBUG nova.compute.manager [req-e050b23a-640e-4f9d-8f2f-b23fdf12a4c4 req-9f2295bd-b6f3-49dc-b3ef-f4ab35d949d2 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] No waiting events found dispatching network-vif-plugged-41768d32-20ce-414e-9c85-be16e03d795a {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1099.982563] env[61243]: WARNING nova.compute.manager [req-e050b23a-640e-4f9d-8f2f-b23fdf12a4c4 req-9f2295bd-b6f3-49dc-b3ef-f4ab35d949d2 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Received unexpected event network-vif-plugged-41768d32-20ce-414e-9c85-be16e03d795a for instance with vm_state building and task_state spawning. [ 1099.982731] env[61243]: DEBUG nova.compute.manager [req-e050b23a-640e-4f9d-8f2f-b23fdf12a4c4 req-9f2295bd-b6f3-49dc-b3ef-f4ab35d949d2 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Received event network-changed-41768d32-20ce-414e-9c85-be16e03d795a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1099.982889] env[61243]: DEBUG nova.compute.manager [req-e050b23a-640e-4f9d-8f2f-b23fdf12a4c4 req-9f2295bd-b6f3-49dc-b3ef-f4ab35d949d2 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Refreshing instance network info cache due to event network-changed-41768d32-20ce-414e-9c85-be16e03d795a. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1099.983076] env[61243]: DEBUG oslo_concurrency.lockutils [req-e050b23a-640e-4f9d-8f2f-b23fdf12a4c4 req-9f2295bd-b6f3-49dc-b3ef-f4ab35d949d2 service nova] Acquiring lock "refresh_cache-91b14a22-811a-4151-9769-9c684ec993db" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.001819] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Instance VIF info [] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1100.007432] env[61243]: DEBUG oslo.service.loopingcall [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1100.010422] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] VM already powered off {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1100.010605] env[61243]: DEBUG nova.compute.manager [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1100.011214] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1100.012230] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44768723-89a5-4bc3-a6a2-948ab2acd091 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.014684] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1fd3e860-fdd1-47ee-9704-d19bc654b371 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.031588] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52af3d8d-015c-fecb-2cb5-6b5237b3f1bf, 'name': SearchDatastore_Task, 'duration_secs': 0.010501} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.032674] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.032964] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] d24f3683-169b-46a8-9387-b6fc88b654bd/d24f3683-169b-46a8-9387-b6fc88b654bd.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1100.033242] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ec24e682-4374-47e3-9c03-fd6877d29e69 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.037106] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.037314] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.041020] env[61243]: DEBUG nova.network.neutron [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1100.041020] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1100.041020] env[61243]: value = "task-1339544" [ 1100.041020] env[61243]: _type = "Task" [ 1100.041020] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.047823] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1100.047823] env[61243]: value = "task-1339545" [ 1100.047823] env[61243]: _type = "Task" [ 1100.047823] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.051628] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339544, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.059554] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339545, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.128386] env[61243]: DEBUG nova.network.neutron [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Updating instance_info_cache with network_info: [{"id": "41768d32-20ce-414e-9c85-be16e03d795a", "address": "fa:16:3e:3c:5e:26", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41768d32-20", "ovs_interfaceid": "41768d32-20ce-414e-9c85-be16e03d795a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.371309] env[61243]: DEBUG oslo_vmware.api [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.372941] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: f77c2610-fa49-487c-9fb4-3b112f800daf] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1100.399918] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.618s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.520103] env[61243]: INFO nova.compute.manager [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Took 18.91 seconds to build instance. [ 1100.555634] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339544, 'name': CreateVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.565771] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339545, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.631326] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "refresh_cache-91b14a22-811a-4151-9769-9c684ec993db" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.631662] env[61243]: DEBUG nova.compute.manager [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Instance network_info: |[{"id": "41768d32-20ce-414e-9c85-be16e03d795a", "address": "fa:16:3e:3c:5e:26", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41768d32-20", "ovs_interfaceid": "41768d32-20ce-414e-9c85-be16e03d795a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1100.631991] env[61243]: DEBUG oslo_concurrency.lockutils [req-e050b23a-640e-4f9d-8f2f-b23fdf12a4c4 req-9f2295bd-b6f3-49dc-b3ef-f4ab35d949d2 service nova] Acquired lock "refresh_cache-91b14a22-811a-4151-9769-9c684ec993db" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.632202] env[61243]: DEBUG nova.network.neutron [req-e050b23a-640e-4f9d-8f2f-b23fdf12a4c4 req-9f2295bd-b6f3-49dc-b3ef-f4ab35d949d2 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Refreshing network info cache for port 41768d32-20ce-414e-9c85-be16e03d795a {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1100.633507] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:5e:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ff3ecd2f-0b10-4faf-a512-fd7a20c28df1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '41768d32-20ce-414e-9c85-be16e03d795a', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1100.640675] env[61243]: DEBUG oslo.service.loopingcall [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1100.643959] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1100.644464] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-170c0c7b-9450-4c71-958b-1f4b2f825d73 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.666578] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1100.666578] env[61243]: value = "task-1339546" [ 1100.666578] env[61243]: _type = "Task" [ 1100.666578] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.674174] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339546, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.866557] env[61243]: DEBUG nova.network.neutron [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Updating instance_info_cache with network_info: [{"id": "00f9d661-7722-4165-8933-c025ac1f938a", "address": "fa:16:3e:e9:b9:d7", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00f9d661-77", "ovs_interfaceid": "00f9d661-7722-4165-8933-c025ac1f938a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.875824] env[61243]: DEBUG oslo_vmware.api [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.879163] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: b0b5b215-d7ff-4960-9d39-cf792d8d4656] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1100.930120] env[61243]: DEBUG nova.network.neutron [req-e050b23a-640e-4f9d-8f2f-b23fdf12a4c4 req-9f2295bd-b6f3-49dc-b3ef-f4ab35d949d2 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Updated VIF entry in instance network info cache for port 41768d32-20ce-414e-9c85-be16e03d795a. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1100.930515] env[61243]: DEBUG nova.network.neutron [req-e050b23a-640e-4f9d-8f2f-b23fdf12a4c4 req-9f2295bd-b6f3-49dc-b3ef-f4ab35d949d2 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Updating instance_info_cache with network_info: [{"id": "41768d32-20ce-414e-9c85-be16e03d795a", "address": "fa:16:3e:3c:5e:26", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41768d32-20", "ovs_interfaceid": "41768d32-20ce-414e-9c85-be16e03d795a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.022522] env[61243]: DEBUG oslo_concurrency.lockutils [None req-bd114784-2245-4089-b78b-86184cb5eacc tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "3d6acdec-ec3a-46b6-858a-135a734ea8ae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.416s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.054290] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339544, 'name': CreateVM_Task, 'duration_secs': 0.528917} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.054446] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1101.054922] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.055117] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.055459] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1101.055759] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8116c33-1b64-47b6-8ad9-214ea29bb189 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.063279] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1101.063279] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cf3836-3487-376e-7da2-d7f77ffd6e55" [ 1101.063279] env[61243]: _type = "Task" [ 1101.063279] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.066740] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339545, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.584413} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.069606] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] d24f3683-169b-46a8-9387-b6fc88b654bd/d24f3683-169b-46a8-9387-b6fc88b654bd.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1101.069827] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1101.070373] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9f96a3e8-4836-4d18-b7e2-061ecb29c7b6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.078127] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52cf3836-3487-376e-7da2-d7f77ffd6e55, 'name': SearchDatastore_Task, 'duration_secs': 0.010498} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.079299] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.079543] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1101.079811] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.079990] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.080197] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1101.080506] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1101.080506] env[61243]: value = "task-1339547" [ 1101.080506] env[61243]: _type = "Task" [ 1101.080506] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.080725] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b0496de-4a1c-4a31-ac85-fe9c6b6e5345 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.090192] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339547, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.091283] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1101.091443] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1101.092142] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-024bdd97-b58f-440b-b08f-b8e36ff7bc7e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.097142] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1101.097142] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]520b4fc8-6e6b-fbc1-a1c6-b52e244879e0" [ 1101.097142] env[61243]: _type = "Task" [ 1101.097142] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.104785] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]520b4fc8-6e6b-fbc1-a1c6-b52e244879e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.176382] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339546, 'name': CreateVM_Task} progress is 25%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.368980] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Releasing lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.374256] env[61243]: DEBUG oslo_vmware.api [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.382317] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 09587d68-f7fd-4648-bd5c-e53b1b6464e8] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.433097] env[61243]: DEBUG oslo_concurrency.lockutils [req-e050b23a-640e-4f9d-8f2f-b23fdf12a4c4 req-9f2295bd-b6f3-49dc-b3ef-f4ab35d949d2 service nova] Releasing lock "refresh_cache-91b14a22-811a-4151-9769-9c684ec993db" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.593418] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339547, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067733} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.597105] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1101.598392] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8823ca9-2846-4221-9f35-12fb00128e9e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.611819] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]520b4fc8-6e6b-fbc1-a1c6-b52e244879e0, 'name': SearchDatastore_Task, 'duration_secs': 0.009216} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.629603] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] d24f3683-169b-46a8-9387-b6fc88b654bd/d24f3683-169b-46a8-9387-b6fc88b654bd.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1101.629928] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abcea856-ed0b-406e-ad8b-453813f9f18f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.632360] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3ce4598-7627-46b5-987c-a6e4c689a699 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.651976] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1101.651976] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d7b83c-1f96-0db9-1f62-637d2100309f" [ 1101.651976] env[61243]: _type = "Task" [ 1101.651976] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.656185] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1101.656185] env[61243]: value = "task-1339548" [ 1101.656185] env[61243]: _type = "Task" [ 1101.656185] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.663325] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d7b83c-1f96-0db9-1f62-637d2100309f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.666507] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339548, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.676008] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339546, 'name': CreateVM_Task, 'duration_secs': 0.802729} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.676566] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1101.677295] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.677295] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.677553] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1101.678573] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75307f8b-cdb8-49ad-8fc8-e2a4344527ef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.684764] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1101.684764] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a814ef-69ca-f92c-d395-511539c14638" [ 1101.684764] env[61243]: _type = "Task" [ 1101.684764] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.693432] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a814ef-69ca-f92c-d395-511539c14638, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.722451] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1101.723427] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9db03ed-a3c8-4f6a-b462-1dcefdb87d41 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.732145] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1101.732433] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ce45ab12-4b88-4a80-b926-34a059866634 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.809358] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1101.809602] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1101.809795] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Deleting the datastore file [datastore2] 486ebf65-4376-424c-9deb-f1ef70e44be3 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1101.810097] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-053d1b74-2fc5-40a8-adb9-80de60c8cde2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.818482] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1101.818482] env[61243]: value = "task-1339550" [ 1101.818482] env[61243]: _type = "Task" [ 1101.818482] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.826680] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339550, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.873800] env[61243]: DEBUG oslo_vmware.api [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.885485] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 5c063b4a-3fbb-4222-ab6d-7d27b6b99084] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.945828] env[61243]: INFO nova.compute.manager [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Swapping old allocation on dict_keys(['8f3c282f-58fe-4c5d-80db-5a142cf023e1']) held by migration c6475dac-275e-43dc-b28b-6d7f0683bc1a for instance [ 1101.969707] env[61243]: DEBUG nova.scheduler.client.report [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Overwriting current allocation {'allocations': {'8f3c282f-58fe-4c5d-80db-5a142cf023e1': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 162}}, 'project_id': '66e2ba61b3354ef1b496232264d27600', 'user_id': 'ecfff92a554145c0bd024e59fb3d6f52', 'consumer_generation': 1} on consumer 60fa2863-f0aa-4798-8e5f-e93df47f93e8 {{(pid=61243) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1102.011611] env[61243]: DEBUG nova.compute.manager [req-668610ae-0490-45c2-b053-3e91d31ba009 req-57d70c64-b918-48a1-8522-d5aed9209c5f service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Received event network-vif-unplugged-00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1102.011709] env[61243]: DEBUG oslo_concurrency.lockutils [req-668610ae-0490-45c2-b053-3e91d31ba009 req-57d70c64-b918-48a1-8522-d5aed9209c5f service nova] Acquiring lock "486ebf65-4376-424c-9deb-f1ef70e44be3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.017383] env[61243]: DEBUG oslo_concurrency.lockutils [req-668610ae-0490-45c2-b053-3e91d31ba009 req-57d70c64-b918-48a1-8522-d5aed9209c5f service nova] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.017383] env[61243]: DEBUG oslo_concurrency.lockutils [req-668610ae-0490-45c2-b053-3e91d31ba009 req-57d70c64-b918-48a1-8522-d5aed9209c5f service nova] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.017383] env[61243]: DEBUG nova.compute.manager [req-668610ae-0490-45c2-b053-3e91d31ba009 req-57d70c64-b918-48a1-8522-d5aed9209c5f service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] No waiting events found dispatching network-vif-unplugged-00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1102.017383] env[61243]: WARNING nova.compute.manager [req-668610ae-0490-45c2-b053-3e91d31ba009 req-57d70c64-b918-48a1-8522-d5aed9209c5f service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Received unexpected event network-vif-unplugged-00f9d661-7722-4165-8933-c025ac1f938a for instance with vm_state shelved and task_state shelving_offloading. [ 1102.017383] env[61243]: DEBUG nova.compute.manager [req-668610ae-0490-45c2-b053-3e91d31ba009 req-57d70c64-b918-48a1-8522-d5aed9209c5f service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Received event network-changed-00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1102.017383] env[61243]: DEBUG nova.compute.manager [req-668610ae-0490-45c2-b053-3e91d31ba009 req-57d70c64-b918-48a1-8522-d5aed9209c5f service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Refreshing instance network info cache due to event network-changed-00f9d661-7722-4165-8933-c025ac1f938a. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1102.017383] env[61243]: DEBUG oslo_concurrency.lockutils [req-668610ae-0490-45c2-b053-3e91d31ba009 req-57d70c64-b918-48a1-8522-d5aed9209c5f service nova] Acquiring lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.017383] env[61243]: DEBUG oslo_concurrency.lockutils [req-668610ae-0490-45c2-b053-3e91d31ba009 req-57d70c64-b918-48a1-8522-d5aed9209c5f service nova] Acquired lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.017383] env[61243]: DEBUG nova.network.neutron [req-668610ae-0490-45c2-b053-3e91d31ba009 req-57d70c64-b918-48a1-8522-d5aed9209c5f service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Refreshing network info cache for port 00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1102.061323] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.061513] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.061700] env[61243]: DEBUG nova.network.neutron [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1102.163935] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52d7b83c-1f96-0db9-1f62-637d2100309f, 'name': SearchDatastore_Task, 'duration_secs': 0.034701} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.164673] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1102.165010] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 6786bf24-f760-4f9b-947a-25d19ae08253/6786bf24-f760-4f9b-947a-25d19ae08253.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1102.165301] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-43732a70-38a3-46a5-a604-af5439aae8f1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.170331] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339548, 'name': ReconfigVM_Task, 'duration_secs': 0.289847} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.170972] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Reconfigured VM instance instance-0000006e to attach disk [datastore2] d24f3683-169b-46a8-9387-b6fc88b654bd/d24f3683-169b-46a8-9387-b6fc88b654bd.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1102.172044] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eaa5c304-750c-4b30-bd09-420baad58584 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.175995] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1102.175995] env[61243]: value = "task-1339551" [ 1102.175995] env[61243]: _type = "Task" [ 1102.175995] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.180165] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1102.180165] env[61243]: value = "task-1339552" [ 1102.180165] env[61243]: _type = "Task" [ 1102.180165] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.187295] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339551, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.195396] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339552, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.199022] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52a814ef-69ca-f92c-d395-511539c14638, 'name': SearchDatastore_Task, 'duration_secs': 0.014138} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.199359] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1102.199632] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1102.199878] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.200048] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.200301] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1102.200896] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-83462b77-aab0-4963-9f2b-a6a10763e462 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.210887] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1102.210984] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1102.212119] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca7c9daf-1d18-4c07-a3f7-55937c08271b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.217928] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1102.217928] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527364b2-c1b7-ee71-79e6-3f93157c6b3b" [ 1102.217928] env[61243]: _type = "Task" [ 1102.217928] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.228129] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527364b2-c1b7-ee71-79e6-3f93157c6b3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.328696] env[61243]: DEBUG oslo_vmware.api [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339550, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.332745} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.329352] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1102.329556] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1102.329737] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1102.346548] env[61243]: INFO nova.scheduler.client.report [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Deleted allocations for instance 486ebf65-4376-424c-9deb-f1ef70e44be3 [ 1102.373990] env[61243]: DEBUG oslo_vmware.api [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.389209] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: d46fb6dc-cdd9-4688-b5e7-9efde0f770c9] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.504391] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "3d6acdec-ec3a-46b6-858a-135a734ea8ae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.504761] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "3d6acdec-ec3a-46b6-858a-135a734ea8ae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.505034] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "3d6acdec-ec3a-46b6-858a-135a734ea8ae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.505295] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "3d6acdec-ec3a-46b6-858a-135a734ea8ae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.505512] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "3d6acdec-ec3a-46b6-858a-135a734ea8ae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.508165] env[61243]: INFO nova.compute.manager [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Terminating instance [ 1102.510453] env[61243]: DEBUG nova.compute.manager [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1102.510668] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1102.511635] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0284e42f-325e-4570-b90c-d5ec29541661 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.521860] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1102.522592] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f31911be-85f0-4c3e-b7f0-aea79a1a20ca {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.531797] env[61243]: DEBUG oslo_vmware.api [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1102.531797] env[61243]: value = "task-1339553" [ 1102.531797] env[61243]: _type = "Task" [ 1102.531797] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.543210] env[61243]: DEBUG oslo_vmware.api [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339553, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.694132] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339552, 'name': Rename_Task, 'duration_secs': 0.153591} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.697497] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1102.697822] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339551, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.698074] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-604fa317-1d6b-4a1f-8db8-11725e685f45 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.708141] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1102.708141] env[61243]: value = "task-1339554" [ 1102.708141] env[61243]: _type = "Task" [ 1102.708141] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.721202] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339554, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.732032] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527364b2-c1b7-ee71-79e6-3f93157c6b3b, 'name': SearchDatastore_Task, 'duration_secs': 0.011543} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.733114] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df5c6e96-c61d-4fc3-b285-fc6b7113ba22 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.740508] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1102.740508] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e1b969-d328-61a8-4564-fd4ba5ec9c11" [ 1102.740508] env[61243]: _type = "Task" [ 1102.740508] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.751832] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e1b969-d328-61a8-4564-fd4ba5ec9c11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.850981] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.851272] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.851502] env[61243]: DEBUG nova.objects.instance [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lazy-loading 'resources' on Instance uuid 486ebf65-4376-424c-9deb-f1ef70e44be3 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.874361] env[61243]: DEBUG oslo_vmware.api [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.893398] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: ebe8cc8e-cc35-4ec4-bed7-d55df72a2991] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.942420] env[61243]: DEBUG nova.network.neutron [req-668610ae-0490-45c2-b053-3e91d31ba009 req-57d70c64-b918-48a1-8522-d5aed9209c5f service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Updated VIF entry in instance network info cache for port 00f9d661-7722-4165-8933-c025ac1f938a. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1102.942781] env[61243]: DEBUG nova.network.neutron [req-668610ae-0490-45c2-b053-3e91d31ba009 req-57d70c64-b918-48a1-8522-d5aed9209c5f service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Updating instance_info_cache with network_info: [{"id": "00f9d661-7722-4165-8933-c025ac1f938a", "address": "fa:16:3e:e9:b9:d7", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap00f9d661-77", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.944696] env[61243]: DEBUG nova.network.neutron [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating instance_info_cache with network_info: [{"id": "f2a48425-d436-40c9-a2dc-694e6313da6d", "address": "fa:16:3e:34:84:e2", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2a48425-d4", "ovs_interfaceid": "f2a48425-d436-40c9-a2dc-694e6313da6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.055631] env[61243]: DEBUG oslo_vmware.api [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339553, 'name': PowerOffVM_Task, 'duration_secs': 0.231127} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.055929] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1103.056100] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1103.056412] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f40218cd-558c-4f32-84ba-d6293c4b2abe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.121345] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1103.121345] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1103.121463] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleting the datastore file [datastore1] 3d6acdec-ec3a-46b6-858a-135a734ea8ae {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1103.121740] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a9261763-2a8b-4b23-94a5-2011079cb35e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.135254] env[61243]: DEBUG oslo_vmware.api [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for the task: (returnval){ [ 1103.135254] env[61243]: value = "task-1339556" [ 1103.135254] env[61243]: _type = "Task" [ 1103.135254] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.143855] env[61243]: DEBUG oslo_vmware.api [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339556, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.187134] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339551, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.635334} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.187477] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 6786bf24-f760-4f9b-947a-25d19ae08253/6786bf24-f760-4f9b-947a-25d19ae08253.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1103.187739] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1103.188040] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ca92400f-1774-4693-8b7e-223531b85908 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.196422] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1103.196422] env[61243]: value = "task-1339557" [ 1103.196422] env[61243]: _type = "Task" [ 1103.196422] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.205835] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339557, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.218556] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339554, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.251261] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52e1b969-d328-61a8-4564-fd4ba5ec9c11, 'name': SearchDatastore_Task, 'duration_secs': 0.051205} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.251422] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.251688] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 91b14a22-811a-4151-9769-9c684ec993db/91b14a22-811a-4151-9769-9c684ec993db.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1103.251871] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a2897bbc-ffee-4299-922b-523d2c8bc901 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.260316] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1103.260316] env[61243]: value = "task-1339558" [ 1103.260316] env[61243]: _type = "Task" [ 1103.260316] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.270708] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339558, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.354279] env[61243]: DEBUG nova.objects.instance [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lazy-loading 'numa_topology' on Instance uuid 486ebf65-4376-424c-9deb-f1ef70e44be3 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1103.374983] env[61243]: DEBUG oslo_vmware.api [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339538, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.396168] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: e9a1c8af-7129-4366-bda1-1c3cf40eeab1] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1103.446815] env[61243]: DEBUG oslo_concurrency.lockutils [req-668610ae-0490-45c2-b053-3e91d31ba009 req-57d70c64-b918-48a1-8522-d5aed9209c5f service nova] Releasing lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.447637] env[61243]: DEBUG oslo_concurrency.lockutils [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "refresh_cache-60fa2863-f0aa-4798-8e5f-e93df47f93e8" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.448859] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22f365f3-457b-4e28-820c-65271f6f21fb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.459529] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748dc369-4d74-4ecb-8dc1-dca149c4e02f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.646822] env[61243]: DEBUG oslo_vmware.api [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Task: {'id': task-1339556, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.221778} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.647105] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1103.647243] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1103.647430] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1103.647608] env[61243]: INFO nova.compute.manager [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1103.647864] env[61243]: DEBUG oslo.service.loopingcall [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1103.648077] env[61243]: DEBUG nova.compute.manager [-] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1103.648224] env[61243]: DEBUG nova.network.neutron [-] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1103.707427] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339557, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076648} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.707714] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1103.708565] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98fb91cf-f56e-4d4b-915d-a108dd3f9f74 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.733102] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 6786bf24-f760-4f9b-947a-25d19ae08253/6786bf24-f760-4f9b-947a-25d19ae08253.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1103.736716] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16c2741b-1ad6-4cd9-9e60-e633f3d8f2b8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.752309] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339554, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.758934] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1103.758934] env[61243]: value = "task-1339559" [ 1103.758934] env[61243]: _type = "Task" [ 1103.758934] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.771937] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339559, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.774941] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339558, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.857096] env[61243]: DEBUG nova.objects.base [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Object Instance<486ebf65-4376-424c-9deb-f1ef70e44be3> lazy-loaded attributes: resources,numa_topology {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1103.875969] env[61243]: DEBUG oslo_vmware.api [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339538, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.899328] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: d95a4d61-2bff-4d89-9ab8-f71e558f3409] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1103.987349] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c64694-6dfe-4368-b80e-4f518897e4c0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.996501] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6124cdf8-371a-4e7c-9d21-3e09561bec18 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.041217] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3bf5ba-2b82-4594-9d6c-1a2870108248 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.046202] env[61243]: DEBUG nova.compute.manager [req-0108a70c-fbe3-4549-85fc-48dd77b0a8b2 req-06996f73-6146-48cd-8c15-20a5cc585e50 service nova] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Received event network-vif-deleted-aa211a5b-6d6e-437d-9219-337bfb9be880 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1104.046440] env[61243]: INFO nova.compute.manager [req-0108a70c-fbe3-4549-85fc-48dd77b0a8b2 req-06996f73-6146-48cd-8c15-20a5cc585e50 service nova] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Neutron deleted interface aa211a5b-6d6e-437d-9219-337bfb9be880; detaching it from the instance and deleting it from the info cache [ 1104.046606] env[61243]: DEBUG nova.network.neutron [req-0108a70c-fbe3-4549-85fc-48dd77b0a8b2 req-06996f73-6146-48cd-8c15-20a5cc585e50 service nova] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.059429] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb98442-d164-4923-8754-0c2ab9496c7b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.075676] env[61243]: DEBUG nova.compute.provider_tree [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.224208] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339554, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.274925] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339559, 'name': ReconfigVM_Task, 'duration_secs': 0.320052} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.275154] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339558, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517238} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.275419] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 6786bf24-f760-4f9b-947a-25d19ae08253/6786bf24-f760-4f9b-947a-25d19ae08253.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1104.276089] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 91b14a22-811a-4151-9769-9c684ec993db/91b14a22-811a-4151-9769-9c684ec993db.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1104.276300] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1104.276545] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a09370b1-70fb-4f03-a1fc-80d8cdbffdaf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.278437] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b7fd5768-58fe-40d5-a3c1-1a4de8f312d4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.286117] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1104.286117] env[61243]: value = "task-1339561" [ 1104.286117] env[61243]: _type = "Task" [ 1104.286117] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.291917] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1104.291917] env[61243]: value = "task-1339560" [ 1104.291917] env[61243]: _type = "Task" [ 1104.291917] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.306698] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339560, 'name': Rename_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.377034] env[61243]: DEBUG oslo_vmware.api [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339538, 'name': ReconfigVM_Task, 'duration_secs': 5.894448} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.377386] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.377659] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Reconfigured VM to detach interface {{(pid=61243) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1104.402892] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: f8816b86-10c8-4d64-9659-fbc301893d46] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.423830] env[61243]: DEBUG nova.network.neutron [-] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.551557] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-902e4aa9-0ad6-4172-bef8-c28666d3e9f7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.561848] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcbb8100-c3f1-4923-9b79-46a28b6be2ef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.578403] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1104.579236] env[61243]: DEBUG nova.scheduler.client.report [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1104.582132] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b86b9da-ac0e-4ad3-a40a-ecf334ad3006 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.592610] env[61243]: DEBUG nova.compute.manager [req-0108a70c-fbe3-4549-85fc-48dd77b0a8b2 req-06996f73-6146-48cd-8c15-20a5cc585e50 service nova] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Detach interface failed, port_id=aa211a5b-6d6e-437d-9219-337bfb9be880, reason: Instance 3d6acdec-ec3a-46b6-858a-135a734ea8ae could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1104.594451] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1104.594451] env[61243]: value = "task-1339562" [ 1104.594451] env[61243]: _type = "Task" [ 1104.594451] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.605044] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339562, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.722032] env[61243]: DEBUG oslo_vmware.api [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339554, 'name': PowerOnVM_Task, 'duration_secs': 1.728944} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.722032] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1104.722328] env[61243]: INFO nova.compute.manager [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Took 9.33 seconds to spawn the instance on the hypervisor. [ 1104.722583] env[61243]: DEBUG nova.compute.manager [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1104.723434] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1364f81-fb55-427f-80a7-adb3b468b12e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.799604] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339561, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08984} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.800353] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1104.801164] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e85943f-4975-42e1-9fe9-204f43f84fb6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.807013] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339560, 'name': Rename_Task, 'duration_secs': 0.158701} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.807672] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1104.807938] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e540b2a-d370-4bb7-9b9a-26fd6b382833 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.828200] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 91b14a22-811a-4151-9769-9c684ec993db/91b14a22-811a-4151-9769-9c684ec993db.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1104.830025] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7fc517ce-bdc0-4c65-9407-35928ef00095 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.854728] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1104.854728] env[61243]: value = "task-1339563" [ 1104.854728] env[61243]: _type = "Task" [ 1104.854728] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.862418] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1104.862418] env[61243]: value = "task-1339564" [ 1104.862418] env[61243]: _type = "Task" [ 1104.862418] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.872016] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339563, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.879298] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339564, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.906259] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 6036199d-cda9-42f8-8669-bc099d2b5ffe] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.927081] env[61243]: INFO nova.compute.manager [-] [instance: 3d6acdec-ec3a-46b6-858a-135a734ea8ae] Took 1.28 seconds to deallocate network for instance. [ 1105.094057] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.243s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.106968] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339562, 'name': PowerOffVM_Task, 'duration_secs': 0.441591} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.107335] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1105.108445] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1105.108735] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1105.109023] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1105.109258] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1105.109454] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1105.109821] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1105.110079] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1105.110242] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1105.110386] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1105.110628] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1105.110832] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1105.115946] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0efcf0ce-b8c7-438c-bec6-49ed432bf74e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.133735] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1105.133735] env[61243]: value = "task-1339565" [ 1105.133735] env[61243]: _type = "Task" [ 1105.133735] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.145646] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339565, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.176217] env[61243]: DEBUG oslo_concurrency.lockutils [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "486ebf65-4376-424c-9deb-f1ef70e44be3" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.245893] env[61243]: INFO nova.compute.manager [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Took 21.42 seconds to build instance. [ 1105.365811] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339563, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.375042] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339564, 'name': ReconfigVM_Task, 'duration_secs': 0.365526} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.375405] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 91b14a22-811a-4151-9769-9c684ec993db/91b14a22-811a-4151-9769-9c684ec993db.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1105.376116] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0459c2c0-b83c-443a-bdf6-05e8ea39a4af {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.384106] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1105.384106] env[61243]: value = "task-1339566" [ 1105.384106] env[61243]: _type = "Task" [ 1105.384106] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.393794] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339566, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.410485] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 783d861c-8930-406e-8986-ef995de4c9dd] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1105.433233] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.433528] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.434088] env[61243]: DEBUG nova.objects.instance [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lazy-loading 'resources' on Instance uuid 3d6acdec-ec3a-46b6-858a-135a734ea8ae {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1105.607614] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2fd18419-ea49-4d7b-8699-d9edd394aa44 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.400s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.608533] env[61243]: DEBUG oslo_concurrency.lockutils [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.432s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.608729] env[61243]: INFO nova.compute.manager [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Unshelving [ 1105.644196] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339565, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.696677] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.696852] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquired lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.697058] env[61243]: DEBUG nova.network.neutron [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1105.747986] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9c892358-d244-4240-85a2-afb1034a7323 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d24f3683-169b-46a8-9387-b6fc88b654bd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.927s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.867095] env[61243]: DEBUG oslo_vmware.api [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339563, 'name': PowerOnVM_Task, 'duration_secs': 0.729617} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.867402] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1105.867616] env[61243]: DEBUG nova.compute.manager [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1105.868455] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed01271-39ca-410b-bf4a-15988cd92d25 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.893313] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339566, 'name': Rename_Task, 'duration_secs': 0.150633} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.893667] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1105.893862] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad77034f-55ea-4396-837a-db12b451891a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.900482] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1105.900482] env[61243]: value = "task-1339567" [ 1105.900482] env[61243]: _type = "Task" [ 1105.900482] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.908517] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339567, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.915204] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 8df486c4-d1ce-47b0-81a8-ad9e83eaac7f] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1105.963105] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.963477] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.963781] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.964072] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.964334] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.966950] env[61243]: INFO nova.compute.manager [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Terminating instance [ 1105.973117] env[61243]: DEBUG nova.compute.manager [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1105.973117] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1105.973899] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f53ac4-13be-41dd-a13b-bad2796690ba {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.988883] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1105.989259] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e0c0c34b-9f9e-4187-90dd-a947e2a5965a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.000524] env[61243]: DEBUG oslo_vmware.api [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1106.000524] env[61243]: value = "task-1339568" [ 1106.000524] env[61243]: _type = "Task" [ 1106.000524] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.017968] env[61243]: DEBUG oslo_vmware.api [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339568, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.083581] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2114fb9-71bd-4096-afb0-1d03c99963ab {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.092681] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf57afd-231d-4f88-bf6f-063c6c5e4d7e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.131048] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a71e4a-388d-406e-aa02-829db213d6a1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.145144] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5221062d-61b9-485e-a91c-ed19f20f76e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.155346] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339565, 'name': ReconfigVM_Task, 'duration_secs': 0.532726} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.156500] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feec97af-9d50-470c-b490-142dfce57ee1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.168230] env[61243]: DEBUG nova.compute.provider_tree [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.191952] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1106.192626] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1106.192626] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1106.192766] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1106.192996] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1106.193237] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1106.193502] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1106.193948] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1106.194028] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1106.194263] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1106.194552] env[61243]: DEBUG nova.virt.hardware [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1106.198191] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b60921d-8912-488f-a706-99d948bb6168 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.211767] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1106.211767] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5216393b-fee1-3d42-2cf7-b49359ceedfc" [ 1106.211767] env[61243]: _type = "Task" [ 1106.211767] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.222466] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5216393b-fee1-3d42-2cf7-b49359ceedfc, 'name': SearchDatastore_Task, 'duration_secs': 0.008935} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.228298] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Reconfiguring VM instance instance-00000062 to detach disk 2000 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1106.228815] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9e784cc-1a8f-41d2-85d4-244b732b143a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.249357] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1106.249357] env[61243]: value = "task-1339569" [ 1106.249357] env[61243]: _type = "Task" [ 1106.249357] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.258024] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339569, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.387411] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.410542] env[61243]: DEBUG oslo_vmware.api [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339567, 'name': PowerOnVM_Task, 'duration_secs': 0.469832} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.410837] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1106.411051] env[61243]: INFO nova.compute.manager [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Took 8.68 seconds to spawn the instance on the hypervisor. [ 1106.411240] env[61243]: DEBUG nova.compute.manager [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1106.412007] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f692075-b0d9-4b3e-a67b-4fe7c2fbd86d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.418040] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: ca230425-aaa7-4c32-81d3-379cbce0d41b] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.514255] env[61243]: DEBUG oslo_vmware.api [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339568, 'name': PowerOffVM_Task, 'duration_secs': 0.249128} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.514604] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1106.514803] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1106.515099] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e4d6276b-7271-4438-8594-31d2fe1d9010 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.569391] env[61243]: DEBUG nova.compute.manager [req-f4c79137-84b9-422d-b545-88616e94997e req-64714acc-3961-44c0-bc4e-de03797e210d service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Received event network-changed-1c517f1d-301b-43a0-8ae9-6099f75580ed {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1106.569689] env[61243]: DEBUG nova.compute.manager [req-f4c79137-84b9-422d-b545-88616e94997e req-64714acc-3961-44c0-bc4e-de03797e210d service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Refreshing instance network info cache due to event network-changed-1c517f1d-301b-43a0-8ae9-6099f75580ed. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1106.569980] env[61243]: DEBUG oslo_concurrency.lockutils [req-f4c79137-84b9-422d-b545-88616e94997e req-64714acc-3961-44c0-bc4e-de03797e210d service nova] Acquiring lock "refresh_cache-d24f3683-169b-46a8-9387-b6fc88b654bd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.570218] env[61243]: DEBUG oslo_concurrency.lockutils [req-f4c79137-84b9-422d-b545-88616e94997e req-64714acc-3961-44c0-bc4e-de03797e210d service nova] Acquired lock "refresh_cache-d24f3683-169b-46a8-9387-b6fc88b654bd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.570473] env[61243]: DEBUG nova.network.neutron [req-f4c79137-84b9-422d-b545-88616e94997e req-64714acc-3961-44c0-bc4e-de03797e210d service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Refreshing network info cache for port 1c517f1d-301b-43a0-8ae9-6099f75580ed {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1106.577443] env[61243]: INFO nova.network.neutron [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Port deadb902-8527-45ad-aa9c-bb805228869f from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1106.577853] env[61243]: DEBUG nova.network.neutron [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Updating instance_info_cache with network_info: [{"id": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "address": "fa:16:3e:a4:7d:cc", "network": {"id": "7724b87f-3721-42a5-81ce-27ae98fc99c2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1052802247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "683c027e4adc4cc4a6b8747a8c8c0de9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d2d1a82-6e", "ovs_interfaceid": "5d2d1a82-6eda-413f-9581-9fa9107be4bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.591870] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1106.592069] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1106.592343] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Deleting the datastore file [datastore2] 8279496d-4a90-4962-9cc2-5d5e1ee33ff0 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1106.593191] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-68fa6d1a-a037-49fe-93a4-8210861ef10b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.601149] env[61243]: DEBUG oslo_vmware.api [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1106.601149] env[61243]: value = "task-1339571" [ 1106.601149] env[61243]: _type = "Task" [ 1106.601149] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.611152] env[61243]: DEBUG oslo_vmware.api [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339571, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.664788] env[61243]: DEBUG oslo_concurrency.lockutils [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.674020] env[61243]: DEBUG nova.scheduler.client.report [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1106.707674] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquiring lock "6786bf24-f760-4f9b-947a-25d19ae08253" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.708569] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Lock "6786bf24-f760-4f9b-947a-25d19ae08253" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.709021] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquiring lock "6786bf24-f760-4f9b-947a-25d19ae08253-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.709414] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Lock "6786bf24-f760-4f9b-947a-25d19ae08253-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.709770] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Lock "6786bf24-f760-4f9b-947a-25d19ae08253-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.712048] env[61243]: INFO nova.compute.manager [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Terminating instance [ 1106.713922] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquiring lock "refresh_cache-6786bf24-f760-4f9b-947a-25d19ae08253" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.714209] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquired lock "refresh_cache-6786bf24-f760-4f9b-947a-25d19ae08253" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.714514] env[61243]: DEBUG nova.network.neutron [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1106.759518] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339569, 'name': ReconfigVM_Task, 'duration_secs': 0.433602} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.759797] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Reconfigured VM instance instance-00000062 to detach disk 2000 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1106.760600] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f97ed4d-4549-4b98-aed1-0edcab49e59b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.785346] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 60fa2863-f0aa-4798-8e5f-e93df47f93e8/60fa2863-f0aa-4798-8e5f-e93df47f93e8.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1106.785621] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c6c7775-a303-497a-8e07-9817d65ebfa3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.804733] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1106.804733] env[61243]: value = "task-1339572" [ 1106.804733] env[61243]: _type = "Task" [ 1106.804733] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.812827] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339572, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.925689] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 5daccb48-b9bf-423a-ab1a-5aa26d8bec87] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.936192] env[61243]: INFO nova.compute.manager [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Took 20.39 seconds to build instance. [ 1107.082458] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Releasing lock "refresh_cache-8279496d-4a90-4962-9cc2-5d5e1ee33ff0" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.112093] env[61243]: DEBUG oslo_vmware.api [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339571, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.33101} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.112093] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1107.112093] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1107.112340] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1107.112423] env[61243]: INFO nova.compute.manager [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1107.112669] env[61243]: DEBUG oslo.service.loopingcall [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1107.112924] env[61243]: DEBUG nova.compute.manager [-] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1107.113036] env[61243]: DEBUG nova.network.neutron [-] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1107.178410] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.745s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.180904] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.795s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.182103] env[61243]: DEBUG nova.objects.instance [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61243) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1107.210199] env[61243]: INFO nova.scheduler.client.report [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Deleted allocations for instance 3d6acdec-ec3a-46b6-858a-135a734ea8ae [ 1107.247757] env[61243]: DEBUG nova.network.neutron [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1107.318425] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339572, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.377154] env[61243]: DEBUG nova.network.neutron [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.432801] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: f73a4d39-5478-4135-9be6-e59f3e29788d] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1107.437745] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a7bc7003-98d9-4c45-b5c4-09657da8e8a9 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "91b14a22-811a-4151-9769-9c684ec993db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.903s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.589128] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7341ccf6-810b-4994-8dd9-5e8d5361da7d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "interface-8279496d-4a90-4962-9cc2-5d5e1ee33ff0-deadb902-8527-45ad-aa9c-bb805228869f" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.213s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.698572] env[61243]: DEBUG nova.network.neutron [req-f4c79137-84b9-422d-b545-88616e94997e req-64714acc-3961-44c0-bc4e-de03797e210d service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Updated VIF entry in instance network info cache for port 1c517f1d-301b-43a0-8ae9-6099f75580ed. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1107.698903] env[61243]: DEBUG nova.network.neutron [req-f4c79137-84b9-422d-b545-88616e94997e req-64714acc-3961-44c0-bc4e-de03797e210d service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Updating instance_info_cache with network_info: [{"id": "1c517f1d-301b-43a0-8ae9-6099f75580ed", "address": "fa:16:3e:aa:38:1f", "network": {"id": "523d4cf1-467c-4ea6-80a4-ebbd01016efa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-919325376-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.236", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7651b565d1647eb948fe07faba32a41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c517f1d-30", "ovs_interfaceid": "1c517f1d-301b-43a0-8ae9-6099f75580ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.706431] env[61243]: DEBUG nova.compute.manager [req-f69d0e25-2b56-4e6c-856b-e8da983758ee req-4dbab37b-25a0-4010-9b88-5323863ff662 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Received event network-changed-41768d32-20ce-414e-9c85-be16e03d795a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1107.706607] env[61243]: DEBUG nova.compute.manager [req-f69d0e25-2b56-4e6c-856b-e8da983758ee req-4dbab37b-25a0-4010-9b88-5323863ff662 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Refreshing instance network info cache due to event network-changed-41768d32-20ce-414e-9c85-be16e03d795a. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1107.706827] env[61243]: DEBUG oslo_concurrency.lockutils [req-f69d0e25-2b56-4e6c-856b-e8da983758ee req-4dbab37b-25a0-4010-9b88-5323863ff662 service nova] Acquiring lock "refresh_cache-91b14a22-811a-4151-9769-9c684ec993db" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.706974] env[61243]: DEBUG oslo_concurrency.lockutils [req-f69d0e25-2b56-4e6c-856b-e8da983758ee req-4dbab37b-25a0-4010-9b88-5323863ff662 service nova] Acquired lock "refresh_cache-91b14a22-811a-4151-9769-9c684ec993db" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.707154] env[61243]: DEBUG nova.network.neutron [req-f69d0e25-2b56-4e6c-856b-e8da983758ee req-4dbab37b-25a0-4010-9b88-5323863ff662 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Refreshing network info cache for port 41768d32-20ce-414e-9c85-be16e03d795a {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1107.721043] env[61243]: DEBUG oslo_concurrency.lockutils [None req-c4f07513-df94-4493-9e6e-fffd2b91af81 tempest-ServerDiskConfigTestJSON-2013045445 tempest-ServerDiskConfigTestJSON-2013045445-project-member] Lock "3d6acdec-ec3a-46b6-858a-135a734ea8ae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.216s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.818314] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339572, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.880321] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Releasing lock "refresh_cache-6786bf24-f760-4f9b-947a-25d19ae08253" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.881071] env[61243]: DEBUG nova.compute.manager [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1107.881071] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1107.882678] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a70324-e17e-4080-af3f-04dc8b7953da {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.891497] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1107.892203] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-752b90cb-4f65-4b80-a186-87adf1cd5b70 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.899217] env[61243]: DEBUG oslo_vmware.api [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1107.899217] env[61243]: value = "task-1339573" [ 1107.899217] env[61243]: _type = "Task" [ 1107.899217] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.908054] env[61243]: DEBUG oslo_vmware.api [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339573, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.936779] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: c1a0d242-4eae-4c03-8341-840b41341f17] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1108.194475] env[61243]: DEBUG oslo_concurrency.lockutils [None req-41f934c4-3240-40b7-bc89-f3f86c8fc67c tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.196355] env[61243]: DEBUG oslo_concurrency.lockutils [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.532s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.196903] env[61243]: DEBUG nova.objects.instance [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lazy-loading 'pci_requests' on Instance uuid 486ebf65-4376-424c-9deb-f1ef70e44be3 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1108.201439] env[61243]: DEBUG oslo_concurrency.lockutils [req-f4c79137-84b9-422d-b545-88616e94997e req-64714acc-3961-44c0-bc4e-de03797e210d service nova] Releasing lock "refresh_cache-d24f3683-169b-46a8-9387-b6fc88b654bd" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.324702] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339572, 'name': ReconfigVM_Task, 'duration_secs': 1.383432} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.330482] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 60fa2863-f0aa-4798-8e5f-e93df47f93e8/60fa2863-f0aa-4798-8e5f-e93df47f93e8.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1108.330482] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57addd7e-4bc3-4207-8cea-b5825df327cf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.366338] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452125e9-c84a-427e-b80e-8387aa740ee3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.397072] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295a5583-659b-41b4-a985-5bd79a6aba6e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.420682] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada1b8f7-53fd-4366-b647-461624efa47f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.426760] env[61243]: DEBUG oslo_vmware.api [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339573, 'name': PowerOffVM_Task, 'duration_secs': 0.3532} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.427691] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1108.427691] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1108.427857] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bae85ec6-dda3-43ae-9c5f-17da79479a94 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.431935] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1108.432618] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ba883d59-7cb3-48e7-a112-d31b45f40033 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.441040] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1108.441040] env[61243]: value = "task-1339575" [ 1108.441040] env[61243]: _type = "Task" [ 1108.441040] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.444828] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 316572e0-c007-42cb-aaf0-3a8cfcaf24aa] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1108.453695] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339575, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.457751] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1108.458651] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1108.458651] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Deleting the datastore file [datastore2] 6786bf24-f760-4f9b-947a-25d19ae08253 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1108.458651] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bfd30ab4-bec8-4248-b54b-47547fdced1b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.466770] env[61243]: DEBUG oslo_vmware.api [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for the task: (returnval){ [ 1108.466770] env[61243]: value = "task-1339576" [ 1108.466770] env[61243]: _type = "Task" [ 1108.466770] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.477333] env[61243]: DEBUG oslo_vmware.api [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339576, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.518307] env[61243]: DEBUG nova.network.neutron [req-f69d0e25-2b56-4e6c-856b-e8da983758ee req-4dbab37b-25a0-4010-9b88-5323863ff662 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Updated VIF entry in instance network info cache for port 41768d32-20ce-414e-9c85-be16e03d795a. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1108.518564] env[61243]: DEBUG nova.network.neutron [req-f69d0e25-2b56-4e6c-856b-e8da983758ee req-4dbab37b-25a0-4010-9b88-5323863ff662 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Updating instance_info_cache with network_info: [{"id": "41768d32-20ce-414e-9c85-be16e03d795a", "address": "fa:16:3e:3c:5e:26", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41768d32-20", "ovs_interfaceid": "41768d32-20ce-414e-9c85-be16e03d795a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.599738] env[61243]: DEBUG nova.compute.manager [req-572ff6d6-a6a6-412a-9ab3-073e11852ec1 req-ff2f9afc-3ae1-44af-88ab-083a67712643 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Received event network-vif-deleted-5d2d1a82-6eda-413f-9581-9fa9107be4bd {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1108.600030] env[61243]: INFO nova.compute.manager [req-572ff6d6-a6a6-412a-9ab3-073e11852ec1 req-ff2f9afc-3ae1-44af-88ab-083a67712643 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Neutron deleted interface 5d2d1a82-6eda-413f-9581-9fa9107be4bd; detaching it from the instance and deleting it from the info cache [ 1108.600266] env[61243]: DEBUG nova.network.neutron [req-572ff6d6-a6a6-412a-9ab3-073e11852ec1 req-ff2f9afc-3ae1-44af-88ab-083a67712643 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.701453] env[61243]: DEBUG nova.objects.instance [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lazy-loading 'numa_topology' on Instance uuid 486ebf65-4376-424c-9deb-f1ef70e44be3 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1108.870164] env[61243]: DEBUG nova.network.neutron [-] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.947623] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 36db1ed5-846f-4ad6-8cee-38b73ff00321] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1108.955408] env[61243]: DEBUG oslo_vmware.api [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339575, 'name': PowerOnVM_Task, 'duration_secs': 0.490796} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.956363] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1108.978941] env[61243]: DEBUG oslo_vmware.api [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Task: {'id': task-1339576, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175835} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.980672] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1108.981077] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1108.982016] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1108.982016] env[61243]: INFO nova.compute.manager [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1108.982299] env[61243]: DEBUG oslo.service.loopingcall [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1108.985157] env[61243]: DEBUG nova.compute.manager [-] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1108.985157] env[61243]: DEBUG nova.network.neutron [-] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1109.002624] env[61243]: DEBUG nova.network.neutron [-] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1109.021429] env[61243]: DEBUG oslo_concurrency.lockutils [req-f69d0e25-2b56-4e6c-856b-e8da983758ee req-4dbab37b-25a0-4010-9b88-5323863ff662 service nova] Releasing lock "refresh_cache-91b14a22-811a-4151-9769-9c684ec993db" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.103434] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-54353ea9-6926-4c0b-94b7-420b053dce8c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.118015] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781d232b-36f3-4347-84d4-fda04b56f62d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.156522] env[61243]: DEBUG nova.compute.manager [req-572ff6d6-a6a6-412a-9ab3-073e11852ec1 req-ff2f9afc-3ae1-44af-88ab-083a67712643 service nova] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Detach interface failed, port_id=5d2d1a82-6eda-413f-9581-9fa9107be4bd, reason: Instance 8279496d-4a90-4962-9cc2-5d5e1ee33ff0 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1109.203673] env[61243]: INFO nova.compute.claims [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1109.372753] env[61243]: INFO nova.compute.manager [-] [instance: 8279496d-4a90-4962-9cc2-5d5e1ee33ff0] Took 2.26 seconds to deallocate network for instance. [ 1109.456429] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 1dc8ad6a-3a00-47c6-8985-481e7f1363cd] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1109.504951] env[61243]: DEBUG nova.network.neutron [-] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.881055] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.962550] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 9a66fd05-1880-423a-8f60-4f11f1c75ab3] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1110.007723] env[61243]: INFO nova.compute.manager [-] [instance: 6786bf24-f760-4f9b-947a-25d19ae08253] Took 1.02 seconds to deallocate network for instance. [ 1110.015656] env[61243]: INFO nova.compute.manager [None req-3c1ad575-dbc9-4a27-80ce-92a4cee9c37d tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating instance to original state: 'active' [ 1110.313711] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c61cd61-7b1f-40d2-84fc-6ad02fe065a0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.322265] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5e370c-d4c3-4e41-aa36-0f7ac7413f74 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.352051] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bcd38ba-e846-4005-a3bb-64b35f1672e7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.360183] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0526ca-1b2e-4f15-ac9b-116f2b6ddde5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.374298] env[61243]: DEBUG nova.compute.provider_tree [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1110.468169] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 6f33b303-1ecd-478e-8fa3-33a5f9a5af50] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1110.517044] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.878438] env[61243]: DEBUG nova.scheduler.client.report [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1110.971853] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 96315d9e-4eda-4e3a-af0d-bdc52ab181e3] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1111.329873] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.330517] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.331050] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.331540] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.331974] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.335824] env[61243]: INFO nova.compute.manager [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Terminating instance [ 1111.339921] env[61243]: DEBUG nova.compute.manager [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1111.340302] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1111.340866] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-337aa383-b74b-4341-8126-e25240daee98 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.357541] env[61243]: DEBUG oslo_vmware.api [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1111.357541] env[61243]: value = "task-1339578" [ 1111.357541] env[61243]: _type = "Task" [ 1111.357541] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.371930] env[61243]: DEBUG oslo_vmware.api [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339578, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.382910] env[61243]: DEBUG oslo_concurrency.lockutils [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.187s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.385175] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.504s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.385457] env[61243]: DEBUG nova.objects.instance [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lazy-loading 'resources' on Instance uuid 8279496d-4a90-4962-9cc2-5d5e1ee33ff0 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1111.428788] env[61243]: INFO nova.network.neutron [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Updating port 00f9d661-7722-4165-8933-c025ac1f938a with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1111.475230] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 729d9bd9-1ab0-47bd-9b7a-11412d3dc608] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1111.867870] env[61243]: DEBUG oslo_vmware.api [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339578, 'name': PowerOffVM_Task, 'duration_secs': 0.24506} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.868169] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1111.868426] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Volume detach. Driver type: vmdk {{(pid=61243) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1111.868640] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285791', 'volume_id': '77ec08cb-5d18-4bee-b234-4aca1c5ce999', 'name': 'volume-77ec08cb-5d18-4bee-b234-4aca1c5ce999', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '60fa2863-f0aa-4798-8e5f-e93df47f93e8', 'attached_at': '2024-10-22T18:13:32.000000', 'detached_at': '', 'volume_id': '77ec08cb-5d18-4bee-b234-4aca1c5ce999', 'serial': '77ec08cb-5d18-4bee-b234-4aca1c5ce999'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1111.869423] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dbf4e37-6ad0-4f1a-8d8b-c577ca7df15e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.894255] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f459474-e4b6-45b5-b48c-92895b8a3656 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.902249] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe39e3d-029f-4d65-99d1-440a1681eb8a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.926364] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5acc3b61-3ab4-4461-8d30-2c0c7e4f8eb5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.944959] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] The volume has not been displaced from its original location: [datastore1] volume-77ec08cb-5d18-4bee-b234-4aca1c5ce999/volume-77ec08cb-5d18-4bee-b234-4aca1c5ce999.vmdk. No consolidation needed. {{(pid=61243) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1111.950551] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1111.953503] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbb5d4ae-21f9-469c-88d6-5cb201e58d83 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.973377] env[61243]: DEBUG oslo_vmware.api [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1111.973377] env[61243]: value = "task-1339579" [ 1111.973377] env[61243]: _type = "Task" [ 1111.973377] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.979665] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 99ce6ca3-a478-4ebe-bf1b-42459aacd9ba] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1111.987924] env[61243]: DEBUG oslo_vmware.api [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339579, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.054785] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe683e1-519e-47c3-9616-d8faffc24c51 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.063396] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e144d0-e427-4a61-bad1-f8c8c2eef3c6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.095016] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b7a3c5-456a-47fe-b6fc-b9ab68513132 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.103356] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d052f1b8-b1db-4983-a2d9-11200ef80591 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.118385] env[61243]: DEBUG nova.compute.provider_tree [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1112.484808] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 894eb798-6c9e-47cb-8eb5-4610fe184bc5] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1112.486659] env[61243]: DEBUG oslo_vmware.api [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339579, 'name': ReconfigVM_Task, 'duration_secs': 0.29334} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.487152] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1112.492104] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-577a6304-b1bc-40d1-96f2-93e7601b3b21 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.508864] env[61243]: DEBUG oslo_vmware.api [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1112.508864] env[61243]: value = "task-1339581" [ 1112.508864] env[61243]: _type = "Task" [ 1112.508864] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.519332] env[61243]: DEBUG oslo_vmware.api [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339581, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.622525] env[61243]: DEBUG nova.scheduler.client.report [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1112.839811] env[61243]: DEBUG nova.compute.manager [req-a733fd8b-1c9c-4872-b297-64d4eaffea57 req-f8dad09d-7f10-437c-a7ec-385914347d75 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Received event network-vif-plugged-00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1112.840060] env[61243]: DEBUG oslo_concurrency.lockutils [req-a733fd8b-1c9c-4872-b297-64d4eaffea57 req-f8dad09d-7f10-437c-a7ec-385914347d75 service nova] Acquiring lock "486ebf65-4376-424c-9deb-f1ef70e44be3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.840288] env[61243]: DEBUG oslo_concurrency.lockutils [req-a733fd8b-1c9c-4872-b297-64d4eaffea57 req-f8dad09d-7f10-437c-a7ec-385914347d75 service nova] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.840464] env[61243]: DEBUG oslo_concurrency.lockutils [req-a733fd8b-1c9c-4872-b297-64d4eaffea57 req-f8dad09d-7f10-437c-a7ec-385914347d75 service nova] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.840633] env[61243]: DEBUG nova.compute.manager [req-a733fd8b-1c9c-4872-b297-64d4eaffea57 req-f8dad09d-7f10-437c-a7ec-385914347d75 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] No waiting events found dispatching network-vif-plugged-00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1112.840802] env[61243]: WARNING nova.compute.manager [req-a733fd8b-1c9c-4872-b297-64d4eaffea57 req-f8dad09d-7f10-437c-a7ec-385914347d75 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Received unexpected event network-vif-plugged-00f9d661-7722-4165-8933-c025ac1f938a for instance with vm_state shelved_offloaded and task_state spawning. [ 1112.921228] env[61243]: DEBUG oslo_concurrency.lockutils [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.921452] env[61243]: DEBUG oslo_concurrency.lockutils [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.921649] env[61243]: DEBUG nova.network.neutron [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1112.987908] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 6cee3dd1-c13d-427d-889c-6cd3ae388a31] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1113.019473] env[61243]: DEBUG oslo_vmware.api [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339581, 'name': ReconfigVM_Task, 'duration_secs': 0.194026} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.019780] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285791', 'volume_id': '77ec08cb-5d18-4bee-b234-4aca1c5ce999', 'name': 'volume-77ec08cb-5d18-4bee-b234-4aca1c5ce999', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '60fa2863-f0aa-4798-8e5f-e93df47f93e8', 'attached_at': '2024-10-22T18:13:32.000000', 'detached_at': '', 'volume_id': '77ec08cb-5d18-4bee-b234-4aca1c5ce999', 'serial': '77ec08cb-5d18-4bee-b234-4aca1c5ce999'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1113.020111] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1113.020844] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9057a7ca-59c3-4f88-8085-ae4e8744cfc3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.027406] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1113.027623] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e04cb3cb-37f5-4fe4-a7ad-27937634eee2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.105420] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1113.105690] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1113.105952] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Deleting the datastore file [datastore2] 60fa2863-f0aa-4798-8e5f-e93df47f93e8 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1113.106350] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ab19456b-0e4c-4ebc-b81f-13634c23d347 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.114205] env[61243]: DEBUG oslo_vmware.api [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1113.114205] env[61243]: value = "task-1339583" [ 1113.114205] env[61243]: _type = "Task" [ 1113.114205] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.123435] env[61243]: DEBUG oslo_vmware.api [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339583, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.128208] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.743s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.130346] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.614s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.130684] env[61243]: DEBUG nova.objects.instance [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Lazy-loading 'resources' on Instance uuid 6786bf24-f760-4f9b-947a-25d19ae08253 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1113.147956] env[61243]: INFO nova.scheduler.client.report [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Deleted allocations for instance 8279496d-4a90-4962-9cc2-5d5e1ee33ff0 [ 1113.490612] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 7aa346fa-64bc-49ff-b7f6-fed7a2ce0e00] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1113.624122] env[61243]: DEBUG oslo_vmware.api [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339583, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162173} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.624386] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1113.624575] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1113.624754] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1113.624927] env[61243]: INFO nova.compute.manager [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Took 2.28 seconds to destroy the instance on the hypervisor. [ 1113.625183] env[61243]: DEBUG oslo.service.loopingcall [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1113.625381] env[61243]: DEBUG nova.compute.manager [-] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1113.625476] env[61243]: DEBUG nova.network.neutron [-] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1113.656827] env[61243]: DEBUG nova.network.neutron [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Updating instance_info_cache with network_info: [{"id": "00f9d661-7722-4165-8933-c025ac1f938a", "address": "fa:16:3e:e9:b9:d7", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00f9d661-77", "ovs_interfaceid": "00f9d661-7722-4165-8933-c025ac1f938a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.658394] env[61243]: DEBUG oslo_concurrency.lockutils [None req-9a718d5a-4fb9-4949-9ac8-e0598680dca8 tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "8279496d-4a90-4962-9cc2-5d5e1ee33ff0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.695s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.742298] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1816a2c-b8e9-4805-9dc5-7d188dc43475 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.751278] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff035a34-740b-4b3d-8072-4e48d72942e5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.788376] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4735624-c6bb-4f76-b54e-c0877d22e88c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.797212] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45fa3244-b1ae-4a0b-8ec6-6af02d088bad {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.801275] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.801557] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.801737] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.801922] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.802134] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.805076] env[61243]: INFO nova.compute.manager [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Terminating instance [ 1113.814039] env[61243]: DEBUG nova.compute.provider_tree [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1113.815577] env[61243]: DEBUG nova.compute.manager [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1113.815775] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1113.816811] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06f4d7c-c569-41a6-aa33-edd34538026f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.824741] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1113.825504] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d98c2f8-9c1c-4264-98d6-c687cea0261d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.833317] env[61243]: DEBUG oslo_vmware.api [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1113.833317] env[61243]: value = "task-1339584" [ 1113.833317] env[61243]: _type = "Task" [ 1113.833317] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.842353] env[61243]: DEBUG oslo_vmware.api [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339584, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.994412] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 36e55334-8628-4dd7-a845-f4ae3d8e7ff9] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1114.160946] env[61243]: DEBUG oslo_concurrency.lockutils [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Releasing lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.196825] env[61243]: DEBUG nova.virt.hardware [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='c6376b558d36dc2ff620eb588d66833b',container_format='bare',created_at=2024-10-22T18:13:12Z,direct_url=,disk_format='vmdk',id=713fcb79-3319-4eb2-bdaf-38d7d056d38b,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-519055829-shelved',owner='2465299214be4d5f877f340489215d6d',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-10-22T18:13:26Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1114.197131] env[61243]: DEBUG nova.virt.hardware [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1114.197319] env[61243]: DEBUG nova.virt.hardware [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1114.197515] env[61243]: DEBUG nova.virt.hardware [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1114.197672] env[61243]: DEBUG nova.virt.hardware [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1114.197888] env[61243]: DEBUG nova.virt.hardware [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1114.198198] env[61243]: DEBUG nova.virt.hardware [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1114.198412] env[61243]: DEBUG nova.virt.hardware [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1114.198631] env[61243]: DEBUG nova.virt.hardware [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1114.198814] env[61243]: DEBUG nova.virt.hardware [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1114.199054] env[61243]: DEBUG nova.virt.hardware [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1114.199921] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb136fe-6f56-4491-b3ce-ad2ee40b1a4c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.211377] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b425877a-0bf4-4c20-970f-eb876ba5bbca {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.227283] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e9:b9:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d413776-9a8c-4afd-856f-10dbb062ca95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '00f9d661-7722-4165-8933-c025ac1f938a', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1114.234907] env[61243]: DEBUG oslo.service.loopingcall [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1114.235294] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1114.235513] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8288eae2-3e9d-4d86-9a71-d911e39fae5d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.257954] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1114.257954] env[61243]: value = "task-1339586" [ 1114.257954] env[61243]: _type = "Task" [ 1114.257954] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.268062] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339586, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.321987] env[61243]: DEBUG nova.scheduler.client.report [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1114.348665] env[61243]: DEBUG oslo_vmware.api [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339584, 'name': PowerOffVM_Task, 'duration_secs': 0.475302} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.349104] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1114.349294] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1114.349566] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-29c0bd7c-f0dc-4648-8f45-f19a65d4c079 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.427113] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1114.427455] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1114.427722] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Deleting the datastore file [datastore2] 466f6a0a-9bc2-4709-8625-e6fc65b68f80 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1114.428138] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6a09d8e-46a2-47af-b828-1d983067c85a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.438691] env[61243]: DEBUG oslo_vmware.api [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for the task: (returnval){ [ 1114.438691] env[61243]: value = "task-1339588" [ 1114.438691] env[61243]: _type = "Task" [ 1114.438691] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.449787] env[61243]: DEBUG oslo_vmware.api [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339588, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.499686] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 07ee984d-476e-484d-ba80-0ec2e411faa9] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1114.611597] env[61243]: DEBUG nova.network.neutron [-] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.773188] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339586, 'name': CreateVM_Task, 'duration_secs': 0.445035} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.773556] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1114.774630] env[61243]: DEBUG oslo_concurrency.lockutils [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/713fcb79-3319-4eb2-bdaf-38d7d056d38b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.774934] env[61243]: DEBUG oslo_concurrency.lockutils [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired lock "[datastore1] devstack-image-cache_base/713fcb79-3319-4eb2-bdaf-38d7d056d38b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.775508] env[61243]: DEBUG oslo_concurrency.lockutils [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/713fcb79-3319-4eb2-bdaf-38d7d056d38b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1114.775922] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7556fa4c-c675-4c2d-b49c-a7899cf82c25 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.782574] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1114.782574] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527b5de3-a1af-aa1b-11bc-54ec25bfcf8a" [ 1114.782574] env[61243]: _type = "Task" [ 1114.782574] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.794732] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]527b5de3-a1af-aa1b-11bc-54ec25bfcf8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.828916] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.698s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.856993] env[61243]: INFO nova.scheduler.client.report [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Deleted allocations for instance 6786bf24-f760-4f9b-947a-25d19ae08253 [ 1114.873289] env[61243]: DEBUG nova.compute.manager [req-1bec8e24-9d47-4068-8ea1-744a2f569d33 req-8462e191-0102-40c6-9a8c-7abaad5347e3 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Received event network-changed-00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1114.873383] env[61243]: DEBUG nova.compute.manager [req-1bec8e24-9d47-4068-8ea1-744a2f569d33 req-8462e191-0102-40c6-9a8c-7abaad5347e3 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Refreshing instance network info cache due to event network-changed-00f9d661-7722-4165-8933-c025ac1f938a. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1114.873540] env[61243]: DEBUG oslo_concurrency.lockutils [req-1bec8e24-9d47-4068-8ea1-744a2f569d33 req-8462e191-0102-40c6-9a8c-7abaad5347e3 service nova] Acquiring lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.873691] env[61243]: DEBUG oslo_concurrency.lockutils [req-1bec8e24-9d47-4068-8ea1-744a2f569d33 req-8462e191-0102-40c6-9a8c-7abaad5347e3 service nova] Acquired lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.873858] env[61243]: DEBUG nova.network.neutron [req-1bec8e24-9d47-4068-8ea1-744a2f569d33 req-8462e191-0102-40c6-9a8c-7abaad5347e3 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Refreshing network info cache for port 00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1114.952095] env[61243]: DEBUG oslo_vmware.api [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Task: {'id': task-1339588, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.335561} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.952491] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1114.952691] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1114.952881] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1114.953077] env[61243]: INFO nova.compute.manager [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1114.953339] env[61243]: DEBUG oslo.service.loopingcall [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1114.953547] env[61243]: DEBUG nova.compute.manager [-] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1114.953645] env[61243]: DEBUG nova.network.neutron [-] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1115.002723] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 86179fb0-99df-4b10-a815-c19168e9521e] Instance has had 0 of 5 cleanup attempts {{(pid=61243) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1115.113785] env[61243]: INFO nova.compute.manager [-] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Took 1.49 seconds to deallocate network for instance. [ 1115.294864] env[61243]: DEBUG oslo_concurrency.lockutils [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Releasing lock "[datastore1] devstack-image-cache_base/713fcb79-3319-4eb2-bdaf-38d7d056d38b" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.295144] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Processing image 713fcb79-3319-4eb2-bdaf-38d7d056d38b {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1115.295389] env[61243]: DEBUG oslo_concurrency.lockutils [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/713fcb79-3319-4eb2-bdaf-38d7d056d38b/713fcb79-3319-4eb2-bdaf-38d7d056d38b.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.295545] env[61243]: DEBUG oslo_concurrency.lockutils [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquired lock "[datastore1] devstack-image-cache_base/713fcb79-3319-4eb2-bdaf-38d7d056d38b/713fcb79-3319-4eb2-bdaf-38d7d056d38b.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.295730] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1115.295989] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49d30a18-704e-4613-864a-21e8afad9ef8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.310921] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1115.311153] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1115.311897] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97bc0478-70e5-4f1a-8511-716bae738f47 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.317882] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1115.317882] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5210e2f0-e558-3d27-d982-6f826c769380" [ 1115.317882] env[61243]: _type = "Task" [ 1115.317882] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.326642] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5210e2f0-e558-3d27-d982-6f826c769380, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.364777] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d642ede6-59e3-4207-aa32-aeb42d8f3f24 tempest-ServerShowV257Test-5508083 tempest-ServerShowV257Test-5508083-project-member] Lock "6786bf24-f760-4f9b-947a-25d19ae08253" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.657s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.506249] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1115.506494] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Cleaning up deleted instances with incomplete migration {{(pid=61243) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1115.620314] env[61243]: DEBUG nova.network.neutron [req-1bec8e24-9d47-4068-8ea1-744a2f569d33 req-8462e191-0102-40c6-9a8c-7abaad5347e3 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Updated VIF entry in instance network info cache for port 00f9d661-7722-4165-8933-c025ac1f938a. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1115.621148] env[61243]: DEBUG nova.network.neutron [req-1bec8e24-9d47-4068-8ea1-744a2f569d33 req-8462e191-0102-40c6-9a8c-7abaad5347e3 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Updating instance_info_cache with network_info: [{"id": "00f9d661-7722-4165-8933-c025ac1f938a", "address": "fa:16:3e:e9:b9:d7", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00f9d661-77", "ovs_interfaceid": "00f9d661-7722-4165-8933-c025ac1f938a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.660310] env[61243]: INFO nova.compute.manager [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Took 0.55 seconds to detach 1 volumes for instance. [ 1115.719532] env[61243]: DEBUG nova.network.neutron [-] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.829058] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Preparing fetch location {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1115.829351] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Fetch image to [datastore1] OSTACK_IMG_ea3ea52a-ee74-4bb2-9fc3-c54b263508e5/OSTACK_IMG_ea3ea52a-ee74-4bb2-9fc3-c54b263508e5.vmdk {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1115.829548] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Downloading stream optimized image 713fcb79-3319-4eb2-bdaf-38d7d056d38b to [datastore1] OSTACK_IMG_ea3ea52a-ee74-4bb2-9fc3-c54b263508e5/OSTACK_IMG_ea3ea52a-ee74-4bb2-9fc3-c54b263508e5.vmdk on the data store datastore1 as vApp {{(pid=61243) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1115.829729] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Downloading image file data 713fcb79-3319-4eb2-bdaf-38d7d056d38b to the ESX as VM named 'OSTACK_IMG_ea3ea52a-ee74-4bb2-9fc3-c54b263508e5' {{(pid=61243) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1115.901202] env[61243]: DEBUG oslo_vmware.rw_handles [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1115.901202] env[61243]: value = "resgroup-9" [ 1115.901202] env[61243]: _type = "ResourcePool" [ 1115.901202] env[61243]: }. {{(pid=61243) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1115.901467] env[61243]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-5a518250-90f9-4f00-8498-21802f27b507 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.924444] env[61243]: DEBUG oslo_vmware.rw_handles [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lease: (returnval){ [ 1115.924444] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52914b81-4c7a-7c80-2e87-78f3d784e95b" [ 1115.924444] env[61243]: _type = "HttpNfcLease" [ 1115.924444] env[61243]: } obtained for vApp import into resource pool (val){ [ 1115.924444] env[61243]: value = "resgroup-9" [ 1115.924444] env[61243]: _type = "ResourcePool" [ 1115.924444] env[61243]: }. {{(pid=61243) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1115.924826] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the lease: (returnval){ [ 1115.924826] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52914b81-4c7a-7c80-2e87-78f3d784e95b" [ 1115.924826] env[61243]: _type = "HttpNfcLease" [ 1115.924826] env[61243]: } to be ready. {{(pid=61243) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1115.931875] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1115.931875] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52914b81-4c7a-7c80-2e87-78f3d784e95b" [ 1115.931875] env[61243]: _type = "HttpNfcLease" [ 1115.931875] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1116.009357] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1116.124320] env[61243]: DEBUG oslo_concurrency.lockutils [req-1bec8e24-9d47-4068-8ea1-744a2f569d33 req-8462e191-0102-40c6-9a8c-7abaad5347e3 service nova] Releasing lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.124739] env[61243]: DEBUG nova.compute.manager [req-1bec8e24-9d47-4068-8ea1-744a2f569d33 req-8462e191-0102-40c6-9a8c-7abaad5347e3 service nova] [instance: 60fa2863-f0aa-4798-8e5f-e93df47f93e8] Received event network-vif-deleted-f2a48425-d436-40c9-a2dc-694e6313da6d {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1116.166894] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.167240] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.167483] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.196826] env[61243]: INFO nova.scheduler.client.report [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Deleted allocations for instance 60fa2863-f0aa-4798-8e5f-e93df47f93e8 [ 1116.222771] env[61243]: INFO nova.compute.manager [-] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Took 1.27 seconds to deallocate network for instance. [ 1116.433684] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1116.433684] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52914b81-4c7a-7c80-2e87-78f3d784e95b" [ 1116.433684] env[61243]: _type = "HttpNfcLease" [ 1116.433684] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1116.704908] env[61243]: DEBUG oslo_concurrency.lockutils [None req-7f558ac1-d921-4b0c-95a1-da6b4d0c0a6f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "60fa2863-f0aa-4798-8e5f-e93df47f93e8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.375s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.730619] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.730905] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.731155] env[61243]: DEBUG nova.objects.instance [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lazy-loading 'resources' on Instance uuid 466f6a0a-9bc2-4709-8625-e6fc65b68f80 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.921050] env[61243]: DEBUG nova.compute.manager [req-dce0ef70-1618-463e-a097-a77aebc2c1c5 req-e81d2bc8-1b82-4a77-86f8-8aa747ac0021 service nova] [instance: 466f6a0a-9bc2-4709-8625-e6fc65b68f80] Received event network-vif-deleted-de6bbdd5-c1ee-40d7-a91c-1938f53863a7 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1116.933966] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1116.933966] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52914b81-4c7a-7c80-2e87-78f3d784e95b" [ 1116.933966] env[61243]: _type = "HttpNfcLease" [ 1116.933966] env[61243]: } is ready. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1116.934278] env[61243]: DEBUG oslo_vmware.rw_handles [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1116.934278] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52914b81-4c7a-7c80-2e87-78f3d784e95b" [ 1116.934278] env[61243]: _type = "HttpNfcLease" [ 1116.934278] env[61243]: }. {{(pid=61243) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1116.934993] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a8c9c5-13d0-4e99-8009-4defb2b58156 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.942257] env[61243]: DEBUG oslo_vmware.rw_handles [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c7cbb5-5216-9934-0d13-9affca0f9d99/disk-0.vmdk from lease info. {{(pid=61243) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1116.942442] env[61243]: DEBUG oslo_vmware.rw_handles [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c7cbb5-5216-9934-0d13-9affca0f9d99/disk-0.vmdk. {{(pid=61243) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1117.002961] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.003188] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Starting heal instance info cache {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1117.003346] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Rebuilding the list of instances to heal {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1117.010692] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ca8ce38c-20f1-477d-b499-6c4438534a1b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.327160] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd7df76-7ca6-4b90-a5e9-2e52416069f3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.338908] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577f4a7c-c64f-40cf-91c6-16b425c6dfd9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.377022] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23711458-dda2-406c-81fd-2b3efeaa9ba0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.387678] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-193b9117-4643-4589-8df1-d89de4bc41af {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.403903] env[61243]: DEBUG nova.compute.provider_tree [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1117.550401] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "refresh_cache-3f839aa5-2b9a-4807-b63b-931f74455532" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1117.550560] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquired lock "refresh_cache-3f839aa5-2b9a-4807-b63b-931f74455532" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.550714] env[61243]: DEBUG nova.network.neutron [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Forcefully refreshing network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1117.550874] env[61243]: DEBUG nova.objects.instance [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lazy-loading 'info_cache' on Instance uuid 3f839aa5-2b9a-4807-b63b-931f74455532 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1117.580794] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "440b35e6-eecb-4e20-831d-cf0db0158e81" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.581055] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.907694] env[61243]: DEBUG nova.scheduler.client.report [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1118.083144] env[61243]: DEBUG nova.compute.manager [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1118.175730] env[61243]: DEBUG oslo_vmware.rw_handles [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Completed reading data from the image iterator. {{(pid=61243) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1118.176125] env[61243]: DEBUG oslo_vmware.rw_handles [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c7cbb5-5216-9934-0d13-9affca0f9d99/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1118.178144] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b80c83-505b-46e0-b24e-bd394b77c253 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.190520] env[61243]: DEBUG oslo_vmware.rw_handles [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c7cbb5-5216-9934-0d13-9affca0f9d99/disk-0.vmdk is in state: ready. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1118.190520] env[61243]: DEBUG oslo_vmware.rw_handles [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c7cbb5-5216-9934-0d13-9affca0f9d99/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1118.190520] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-472dfbae-9f9c-453c-b268-5b5740e8f578 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.405612] env[61243]: DEBUG oslo_vmware.rw_handles [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c7cbb5-5216-9934-0d13-9affca0f9d99/disk-0.vmdk. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1118.405882] env[61243]: INFO nova.virt.vmwareapi.images [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Downloaded image file data 713fcb79-3319-4eb2-bdaf-38d7d056d38b [ 1118.406743] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351a42fe-205d-4e56-a180-7b8c2821e2f7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.413141] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.682s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.427475] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a7c07f7e-8e1a-4962-acbb-7e3ac2540ccd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.466830] env[61243]: INFO nova.scheduler.client.report [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Deleted allocations for instance 466f6a0a-9bc2-4709-8625-e6fc65b68f80 [ 1118.490018] env[61243]: INFO nova.virt.vmwareapi.images [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] The imported VM was unregistered [ 1118.492558] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Caching image {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1118.492822] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Creating directory with path [datastore1] devstack-image-cache_base/713fcb79-3319-4eb2-bdaf-38d7d056d38b {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1118.493219] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-66813861-67da-4b25-b34d-e4301711eb92 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.520011] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Created directory with path [datastore1] devstack-image-cache_base/713fcb79-3319-4eb2-bdaf-38d7d056d38b {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1118.520340] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_ea3ea52a-ee74-4bb2-9fc3-c54b263508e5/OSTACK_IMG_ea3ea52a-ee74-4bb2-9fc3-c54b263508e5.vmdk to [datastore1] devstack-image-cache_base/713fcb79-3319-4eb2-bdaf-38d7d056d38b/713fcb79-3319-4eb2-bdaf-38d7d056d38b.vmdk. {{(pid=61243) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1118.520722] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-a618c8b7-fbc2-46c7-8137-857b7db0285d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.529312] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1118.529312] env[61243]: value = "task-1339593" [ 1118.529312] env[61243]: _type = "Task" [ 1118.529312] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.540015] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339593, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.605599] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.605923] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.607594] env[61243]: INFO nova.compute.claims [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1118.975736] env[61243]: DEBUG oslo_concurrency.lockutils [None req-0147e831-4cc9-4b0d-a78f-c97cb0779b2d tempest-AttachInterfacesTestJSON-2006107775 tempest-AttachInterfacesTestJSON-2006107775-project-member] Lock "466f6a0a-9bc2-4709-8625-e6fc65b68f80" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.174s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.042913] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339593, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.388029] env[61243]: DEBUG nova.network.neutron [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Updating instance_info_cache with network_info: [{"id": "a0b324d6-6a73-4956-85c3-38f755003b74", "address": "fa:16:3e:16:b3:46", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0b324d6-6a", "ovs_interfaceid": "a0b324d6-6a73-4956-85c3-38f755003b74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.541197] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339593, 'name': MoveVirtualDisk_Task} progress is 29%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.695553] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30790db3-46b3-4361-8b78-0f9c6cd1b306 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.704401] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae5e6d9-923b-4a1f-83d1-ed9ed6e0e4b2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.738068] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf09ed4-6f81-4022-8670-0bdb67b1c033 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.745731] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464988be-f06e-4686-9349-acc8fc29d1bb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.760047] env[61243]: DEBUG nova.compute.provider_tree [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1119.890827] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Releasing lock "refresh_cache-3f839aa5-2b9a-4807-b63b-931f74455532" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.891056] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Updated the network info_cache for instance {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1119.891287] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1119.891526] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._sync_power_states {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1120.042295] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339593, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.264996] env[61243]: DEBUG nova.scheduler.client.report [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1120.395779] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Getting list of instances from cluster (obj){ [ 1120.395779] env[61243]: value = "domain-c8" [ 1120.395779] env[61243]: _type = "ClusterComputeResource" [ 1120.395779] env[61243]: } {{(pid=61243) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1120.396814] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f940c107-5a1e-4817-b0d8-ffe0fd1e12c1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.412623] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Got total of 4 instances {{(pid=61243) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1120.412846] env[61243]: WARNING nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] While synchronizing instance power states, found 5 instances in the database and 4 instances on the hypervisor. [ 1120.412966] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Triggering sync for uuid 3f839aa5-2b9a-4807-b63b-931f74455532 {{(pid=61243) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1120.413240] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Triggering sync for uuid 486ebf65-4376-424c-9deb-f1ef70e44be3 {{(pid=61243) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1120.413461] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Triggering sync for uuid d24f3683-169b-46a8-9387-b6fc88b654bd {{(pid=61243) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1120.413623] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Triggering sync for uuid 91b14a22-811a-4151-9769-9c684ec993db {{(pid=61243) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1120.413778] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Triggering sync for uuid 440b35e6-eecb-4e20-831d-cf0db0158e81 {{(pid=61243) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1120.414170] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "3f839aa5-2b9a-4807-b63b-931f74455532" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.414433] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "3f839aa5-2b9a-4807-b63b-931f74455532" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.414717] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "486ebf65-4376-424c-9deb-f1ef70e44be3" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.414939] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "d24f3683-169b-46a8-9387-b6fc88b654bd" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.415141] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "d24f3683-169b-46a8-9387-b6fc88b654bd" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.415389] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "91b14a22-811a-4151-9769-9c684ec993db" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.415573] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "91b14a22-811a-4151-9769-9c684ec993db" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.415811] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "440b35e6-eecb-4e20-831d-cf0db0158e81" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.416038] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1120.416181] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61243) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1120.416903] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f028b8a6-d392-4552-869b-f12c89763b3b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.420902] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13e1088-75f5-48c4-88db-534e9f79d3f8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.425263] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dbc949a-eac3-4598-8844-8f429d4d63f3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.543127] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339593, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.770493] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.164s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.771060] env[61243]: DEBUG nova.compute.manager [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1120.943231] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "91b14a22-811a-4151-9769-9c684ec993db" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.526s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.943231] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "3f839aa5-2b9a-4807-b63b-931f74455532" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.528s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.943231] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "d24f3683-169b-46a8-9387-b6fc88b654bd" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.528s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.043459] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339593, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.275923] env[61243]: DEBUG nova.compute.utils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1121.277376] env[61243]: DEBUG nova.compute.manager [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1121.277554] env[61243]: DEBUG nova.network.neutron [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1121.437179] env[61243]: DEBUG nova.policy [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ecfff92a554145c0bd024e59fb3d6f52', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66e2ba61b3354ef1b496232264d27600', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1121.545119] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339593, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.752331] env[61243]: DEBUG nova.network.neutron [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Successfully created port: 723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1121.780913] env[61243]: DEBUG nova.compute.manager [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1122.047388] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339593, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.549724] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339593, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.795087] env[61243]: DEBUG nova.compute.manager [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1122.820774] env[61243]: DEBUG nova.virt.hardware [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T18:01:49Z,direct_url=,disk_format='vmdk',id=6142e969-c114-4502-aa93-c018fb915a86,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fd62d316b05d4dafb8d29066a95a0df5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T18:01:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1122.821046] env[61243]: DEBUG nova.virt.hardware [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1122.821216] env[61243]: DEBUG nova.virt.hardware [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1122.821405] env[61243]: DEBUG nova.virt.hardware [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1122.821557] env[61243]: DEBUG nova.virt.hardware [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1122.821707] env[61243]: DEBUG nova.virt.hardware [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1122.821913] env[61243]: DEBUG nova.virt.hardware [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1122.822143] env[61243]: DEBUG nova.virt.hardware [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1122.822325] env[61243]: DEBUG nova.virt.hardware [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1122.822495] env[61243]: DEBUG nova.virt.hardware [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1122.822673] env[61243]: DEBUG nova.virt.hardware [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1122.823560] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36cdbdcc-7f85-44d7-8111-c5ded71347f3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.831934] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63644bd-f3bb-4969-ae50-4a2c355278fb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.047639] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339593, 'name': MoveVirtualDisk_Task, 'duration_secs': 4.249614} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.047774] env[61243]: INFO nova.virt.vmwareapi.ds_util [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_ea3ea52a-ee74-4bb2-9fc3-c54b263508e5/OSTACK_IMG_ea3ea52a-ee74-4bb2-9fc3-c54b263508e5.vmdk to [datastore1] devstack-image-cache_base/713fcb79-3319-4eb2-bdaf-38d7d056d38b/713fcb79-3319-4eb2-bdaf-38d7d056d38b.vmdk. [ 1123.047898] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Cleaning up location [datastore1] OSTACK_IMG_ea3ea52a-ee74-4bb2-9fc3-c54b263508e5 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1123.048091] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_ea3ea52a-ee74-4bb2-9fc3-c54b263508e5 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1123.048356] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9fe3b730-c624-4f5e-93a3-162f6ce930e9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.056663] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1123.056663] env[61243]: value = "task-1339596" [ 1123.056663] env[61243]: _type = "Task" [ 1123.056663] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.064631] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339596, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.324324] env[61243]: DEBUG nova.compute.manager [req-44fc5161-6926-4796-8065-3cd860e00e2a req-c196b616-3054-4d00-9307-b7cd826a1fc1 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Received event network-vif-plugged-723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1123.324588] env[61243]: DEBUG oslo_concurrency.lockutils [req-44fc5161-6926-4796-8065-3cd860e00e2a req-c196b616-3054-4d00-9307-b7cd826a1fc1 service nova] Acquiring lock "440b35e6-eecb-4e20-831d-cf0db0158e81-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.324849] env[61243]: DEBUG oslo_concurrency.lockutils [req-44fc5161-6926-4796-8065-3cd860e00e2a req-c196b616-3054-4d00-9307-b7cd826a1fc1 service nova] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.325084] env[61243]: DEBUG oslo_concurrency.lockutils [req-44fc5161-6926-4796-8065-3cd860e00e2a req-c196b616-3054-4d00-9307-b7cd826a1fc1 service nova] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.325312] env[61243]: DEBUG nova.compute.manager [req-44fc5161-6926-4796-8065-3cd860e00e2a req-c196b616-3054-4d00-9307-b7cd826a1fc1 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] No waiting events found dispatching network-vif-plugged-723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1123.325532] env[61243]: WARNING nova.compute.manager [req-44fc5161-6926-4796-8065-3cd860e00e2a req-c196b616-3054-4d00-9307-b7cd826a1fc1 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Received unexpected event network-vif-plugged-723d362f-40a3-492a-9a70-f86c847d5d16 for instance with vm_state building and task_state spawning. [ 1123.566759] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339596, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.037383} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.567077] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1123.567215] env[61243]: DEBUG oslo_concurrency.lockutils [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Releasing lock "[datastore1] devstack-image-cache_base/713fcb79-3319-4eb2-bdaf-38d7d056d38b/713fcb79-3319-4eb2-bdaf-38d7d056d38b.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1123.567468] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/713fcb79-3319-4eb2-bdaf-38d7d056d38b/713fcb79-3319-4eb2-bdaf-38d7d056d38b.vmdk to [datastore1] 486ebf65-4376-424c-9deb-f1ef70e44be3/486ebf65-4376-424c-9deb-f1ef70e44be3.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1123.567728] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12315f18-510e-42f5-ad6e-705ffaa26611 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.575362] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1123.575362] env[61243]: value = "task-1339597" [ 1123.575362] env[61243]: _type = "Task" [ 1123.575362] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.584555] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339597, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.888962] env[61243]: DEBUG nova.network.neutron [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Successfully updated port: 723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1123.916828] env[61243]: DEBUG nova.compute.manager [req-080a4dcc-a191-4b7e-b50c-ac7246a77d00 req-638ca740-52e7-431e-9c30-7d889260a3a8 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Received event network-changed-723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1123.917056] env[61243]: DEBUG nova.compute.manager [req-080a4dcc-a191-4b7e-b50c-ac7246a77d00 req-638ca740-52e7-431e-9c30-7d889260a3a8 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Refreshing instance network info cache due to event network-changed-723d362f-40a3-492a-9a70-f86c847d5d16. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1123.917314] env[61243]: DEBUG oslo_concurrency.lockutils [req-080a4dcc-a191-4b7e-b50c-ac7246a77d00 req-638ca740-52e7-431e-9c30-7d889260a3a8 service nova] Acquiring lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.917475] env[61243]: DEBUG oslo_concurrency.lockutils [req-080a4dcc-a191-4b7e-b50c-ac7246a77d00 req-638ca740-52e7-431e-9c30-7d889260a3a8 service nova] Acquired lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.917652] env[61243]: DEBUG nova.network.neutron [req-080a4dcc-a191-4b7e-b50c-ac7246a77d00 req-638ca740-52e7-431e-9c30-7d889260a3a8 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Refreshing network info cache for port 723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1124.087878] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339597, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.400562] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.549260] env[61243]: DEBUG nova.network.neutron [req-080a4dcc-a191-4b7e-b50c-ac7246a77d00 req-638ca740-52e7-431e-9c30-7d889260a3a8 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1124.589152] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339597, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.634132] env[61243]: DEBUG nova.network.neutron [req-080a4dcc-a191-4b7e-b50c-ac7246a77d00 req-638ca740-52e7-431e-9c30-7d889260a3a8 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.087875] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339597, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.137078] env[61243]: DEBUG oslo_concurrency.lockutils [req-080a4dcc-a191-4b7e-b50c-ac7246a77d00 req-638ca740-52e7-431e-9c30-7d889260a3a8 service nova] Releasing lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.137324] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.137416] env[61243]: DEBUG nova.network.neutron [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1125.591946] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339597, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.671254] env[61243]: DEBUG nova.network.neutron [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1125.818233] env[61243]: DEBUG nova.network.neutron [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Updating instance_info_cache with network_info: [{"id": "723d362f-40a3-492a-9a70-f86c847d5d16", "address": "fa:16:3e:dc:f7:98", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap723d362f-40", "ovs_interfaceid": "723d362f-40a3-492a-9a70-f86c847d5d16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.087543] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339597, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.378521} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.087818] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/713fcb79-3319-4eb2-bdaf-38d7d056d38b/713fcb79-3319-4eb2-bdaf-38d7d056d38b.vmdk to [datastore1] 486ebf65-4376-424c-9deb-f1ef70e44be3/486ebf65-4376-424c-9deb-f1ef70e44be3.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1126.088589] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd03c22-87f9-4afe-a58b-b8e2acb0fd0d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.110014] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 486ebf65-4376-424c-9deb-f1ef70e44be3/486ebf65-4376-424c-9deb-f1ef70e44be3.vmdk or device None with type streamOptimized {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1126.110266] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-419a878d-7eb0-46bd-a512-6d59ac386056 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.128870] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1126.128870] env[61243]: value = "task-1339599" [ 1126.128870] env[61243]: _type = "Task" [ 1126.128870] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.136526] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339599, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.322020] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.322497] env[61243]: DEBUG nova.compute.manager [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Instance network_info: |[{"id": "723d362f-40a3-492a-9a70-f86c847d5d16", "address": "fa:16:3e:dc:f7:98", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap723d362f-40", "ovs_interfaceid": "723d362f-40a3-492a-9a70-f86c847d5d16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1126.323072] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:f7:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd7b5f1ef-d4b9-4ec3-b047-17e4cb349d25', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '723d362f-40a3-492a-9a70-f86c847d5d16', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1126.331712] env[61243]: DEBUG oslo.service.loopingcall [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1126.331994] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1126.332291] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c6636b7-2860-4b52-ad8f-ee267bbce1fc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.358274] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1126.358274] env[61243]: value = "task-1339600" [ 1126.358274] env[61243]: _type = "Task" [ 1126.358274] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.366105] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339600, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.640774] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339599, 'name': ReconfigVM_Task, 'duration_secs': 0.290733} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.641224] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 486ebf65-4376-424c-9deb-f1ef70e44be3/486ebf65-4376-424c-9deb-f1ef70e44be3.vmdk or device None with type streamOptimized {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1126.641706] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e16a2ce-2330-4237-bedb-554e77e5b838 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.648949] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1126.648949] env[61243]: value = "task-1339601" [ 1126.648949] env[61243]: _type = "Task" [ 1126.648949] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.656924] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339601, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.868802] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339600, 'name': CreateVM_Task, 'duration_secs': 0.408356} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.868972] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1126.869677] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.869856] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.870202] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1126.870457] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1ba4093-d67c-4430-ae35-4c0ccae98977 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.874800] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1126.874800] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ce89c2-260c-abd5-346c-4f7beeaabca9" [ 1126.874800] env[61243]: _type = "Task" [ 1126.874800] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.882111] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ce89c2-260c-abd5-346c-4f7beeaabca9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.158855] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339601, 'name': Rename_Task, 'duration_secs': 0.384523} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.159129] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1127.159395] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a7babd7a-2a70-4517-acf0-2953b50358d0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.166695] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1127.166695] env[61243]: value = "task-1339602" [ 1127.166695] env[61243]: _type = "Task" [ 1127.166695] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.176204] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339602, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.387059] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52ce89c2-260c-abd5-346c-4f7beeaabca9, 'name': SearchDatastore_Task, 'duration_secs': 0.01223} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.387059] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.387059] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Processing image 6142e969-c114-4502-aa93-c018fb915a86 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1127.387059] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.387363] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.387363] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1127.387568] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cbf49a60-a08a-4b37-bb02-d037db54543c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.396068] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1127.396257] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1127.396979] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8648b91-a359-430d-a73c-6c8b9de07363 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.402182] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1127.402182] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]520a9d73-1d5c-d089-8ec9-8478ac7b3719" [ 1127.402182] env[61243]: _type = "Task" [ 1127.402182] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.410212] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]520a9d73-1d5c-d089-8ec9-8478ac7b3719, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.677337] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339602, 'name': PowerOnVM_Task} progress is 98%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.913640] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]520a9d73-1d5c-d089-8ec9-8478ac7b3719, 'name': SearchDatastore_Task, 'duration_secs': 0.012116} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.914503] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb5d3b27-262c-4367-83d2-d514093c565e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.920178] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1127.920178] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52b83d7d-7c8d-34de-272f-8d704dc1a3e7" [ 1127.920178] env[61243]: _type = "Task" [ 1127.920178] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.928112] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52b83d7d-7c8d-34de-272f-8d704dc1a3e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.178984] env[61243]: DEBUG oslo_vmware.api [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339602, 'name': PowerOnVM_Task, 'duration_secs': 0.650977} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.178984] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1128.284623] env[61243]: DEBUG nova.compute.manager [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1128.285579] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3b9dab-55d4-4e70-943c-4997ec9cb3a5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.413392] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.413775] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.431715] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52b83d7d-7c8d-34de-272f-8d704dc1a3e7, 'name': SearchDatastore_Task, 'duration_secs': 0.010747} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.431964] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.432251] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 440b35e6-eecb-4e20-831d-cf0db0158e81/440b35e6-eecb-4e20-831d-cf0db0158e81.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1128.432536] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07bd15d5-9822-4518-9dba-30151625bb9d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.439617] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1128.439617] env[61243]: value = "task-1339603" [ 1128.439617] env[61243]: _type = "Task" [ 1128.439617] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.447348] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339603, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.804896] env[61243]: DEBUG oslo_concurrency.lockutils [None req-25419d2d-9d5d-481f-a769-b005a4d4483f tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 23.196s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.805585] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 8.391s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.805585] env[61243]: INFO nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] During sync_power_state the instance has a pending task (spawning). Skip. [ 1128.805772] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.916164] env[61243]: DEBUG nova.compute.manager [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Starting instance... {{(pid=61243) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1128.950473] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339603, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.438455] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.438717] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.440296] env[61243]: INFO nova.compute.claims [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1129.452217] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339603, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51266} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.452795] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6142e969-c114-4502-aa93-c018fb915a86/6142e969-c114-4502-aa93-c018fb915a86.vmdk to [datastore2] 440b35e6-eecb-4e20-831d-cf0db0158e81/440b35e6-eecb-4e20-831d-cf0db0158e81.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1129.452795] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Extending root virtual disk to 1048576 {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1129.452965] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c459e744-df39-4b19-ba03-663e8cd7e519 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.460059] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1129.460059] env[61243]: value = "task-1339604" [ 1129.460059] env[61243]: _type = "Task" [ 1129.460059] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.468445] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339604, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.970042] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339604, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069167} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.970327] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Extended root virtual disk {{(pid=61243) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1129.970983] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e32aa4b-e6b2-482b-ad9a-1c8d709be98a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.992623] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 440b35e6-eecb-4e20-831d-cf0db0158e81/440b35e6-eecb-4e20-831d-cf0db0158e81.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1129.993110] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41bfdafd-36b7-434a-92df-b3bd3322acf0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.012014] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1130.012014] env[61243]: value = "task-1339605" [ 1130.012014] env[61243]: _type = "Task" [ 1130.012014] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.019480] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339605, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.524404] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339605, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.545674] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a108475-db54-4aa9-9fca-e51d700c2f2f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.553448] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1efe0867-ddcd-4a90-872f-032790b78bae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.584574] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23b7afc-48af-4887-b556-f73f426849f6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.592371] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63559b8f-24c9-4a89-8b03-62af9e318f60 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.606804] env[61243]: DEBUG nova.compute.provider_tree [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1131.023197] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339605, 'name': ReconfigVM_Task, 'duration_secs': 0.519155} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.023509] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 440b35e6-eecb-4e20-831d-cf0db0158e81/440b35e6-eecb-4e20-831d-cf0db0158e81.vmdk or device None with type sparse {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1131.024143] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e6407519-d0d0-48d8-b536-214192726f3a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.030999] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1131.030999] env[61243]: value = "task-1339606" [ 1131.030999] env[61243]: _type = "Task" [ 1131.030999] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.038568] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339606, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.109881] env[61243]: DEBUG nova.scheduler.client.report [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1131.541406] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339606, 'name': Rename_Task, 'duration_secs': 0.134937} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.541762] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1131.542059] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75ce125f-dc3a-4293-97a5-6b02de991140 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.548478] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1131.548478] env[61243]: value = "task-1339607" [ 1131.548478] env[61243]: _type = "Task" [ 1131.548478] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.555863] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339607, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.614953] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.176s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.615378] env[61243]: DEBUG nova.compute.manager [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Start building networks asynchronously for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1132.058478] env[61243]: DEBUG oslo_vmware.api [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339607, 'name': PowerOnVM_Task, 'duration_secs': 0.44228} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.058825] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1132.059076] env[61243]: INFO nova.compute.manager [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Took 9.27 seconds to spawn the instance on the hypervisor. [ 1132.059271] env[61243]: DEBUG nova.compute.manager [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1132.060106] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32682062-a794-433e-9c8c-feb0dc8c43e4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.121063] env[61243]: DEBUG nova.compute.utils [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1132.121690] env[61243]: DEBUG nova.compute.manager [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Allocating IP information in the background. {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1132.121920] env[61243]: DEBUG nova.network.neutron [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] allocate_for_instance() {{(pid=61243) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1132.160230] env[61243]: DEBUG nova.policy [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b9748805e2c48e4a8fd0354e916aa0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '324ed556b6f5419fa0376a7cfc508827', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61243) authorize /opt/stack/nova/nova/policy.py:201}} [ 1132.417345] env[61243]: DEBUG nova.network.neutron [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Successfully created port: 2168e525-2705-4de2-9879-ac1a43eb9f84 {{(pid=61243) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1132.577513] env[61243]: INFO nova.compute.manager [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Took 13.99 seconds to build instance. [ 1132.625824] env[61243]: DEBUG nova.compute.manager [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Start building block device mappings for instance. {{(pid=61243) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1132.838619] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1133.080030] env[61243]: DEBUG oslo_concurrency.lockutils [None req-ecd6b10a-d7eb-43b8-b227-0d2572859274 tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.499s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.080418] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 12.664s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.081410] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5b4f96-5488-46e7-8785-a6ff94ab56b4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.137252] env[61243]: INFO nova.virt.block_device [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Booting with volume d7581ea4-bb02-421e-a48c-c34488368a12 at /dev/sda [ 1133.154685] env[61243]: DEBUG nova.compute.manager [req-82a2990e-cd5c-4ef9-a003-a4ccb9918d7e req-0e83c0ff-8f35-4128-ac4f-ba245f98fde9 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Received event network-changed-723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1133.154990] env[61243]: DEBUG nova.compute.manager [req-82a2990e-cd5c-4ef9-a003-a4ccb9918d7e req-0e83c0ff-8f35-4128-ac4f-ba245f98fde9 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Refreshing instance network info cache due to event network-changed-723d362f-40a3-492a-9a70-f86c847d5d16. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1133.155254] env[61243]: DEBUG oslo_concurrency.lockutils [req-82a2990e-cd5c-4ef9-a003-a4ccb9918d7e req-0e83c0ff-8f35-4128-ac4f-ba245f98fde9 service nova] Acquiring lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.155412] env[61243]: DEBUG oslo_concurrency.lockutils [req-82a2990e-cd5c-4ef9-a003-a4ccb9918d7e req-0e83c0ff-8f35-4128-ac4f-ba245f98fde9 service nova] Acquired lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.155654] env[61243]: DEBUG nova.network.neutron [req-82a2990e-cd5c-4ef9-a003-a4ccb9918d7e req-0e83c0ff-8f35-4128-ac4f-ba245f98fde9 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Refreshing network info cache for port 723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1133.177729] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1e3ad28b-6d06-4668-8619-5864d26b0bbc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.188169] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-defeeeba-1072-4485-976f-90a71469d2a8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.216795] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-56bc8fa9-0872-4686-8159-20a1090496a5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.226294] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909d986a-977c-4e74-b15c-59e0cfadc718 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.257103] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ffb4c13-2790-47cd-9bad-65acfcf4464f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.264845] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06514297-e54e-4c11-98d5-896324d2291d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.279621] env[61243]: DEBUG nova.virt.block_device [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updating existing volume attachment record: 13fe082d-ccef-455b-b56e-09f416feca55 {{(pid=61243) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1133.302115] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1133.592250] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.512s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.996096] env[61243]: DEBUG nova.network.neutron [req-82a2990e-cd5c-4ef9-a003-a4ccb9918d7e req-0e83c0ff-8f35-4128-ac4f-ba245f98fde9 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Updated VIF entry in instance network info cache for port 723d362f-40a3-492a-9a70-f86c847d5d16. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1133.996573] env[61243]: DEBUG nova.network.neutron [req-82a2990e-cd5c-4ef9-a003-a4ccb9918d7e req-0e83c0ff-8f35-4128-ac4f-ba245f98fde9 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Updating instance_info_cache with network_info: [{"id": "723d362f-40a3-492a-9a70-f86c847d5d16", "address": "fa:16:3e:dc:f7:98", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap723d362f-40", "ovs_interfaceid": "723d362f-40a3-492a-9a70-f86c847d5d16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.055132] env[61243]: DEBUG nova.network.neutron [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Successfully updated port: 2168e525-2705-4de2-9879-ac1a43eb9f84 {{(pid=61243) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1134.302240] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager.update_available_resource {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.499871] env[61243]: DEBUG oslo_concurrency.lockutils [req-82a2990e-cd5c-4ef9-a003-a4ccb9918d7e req-0e83c0ff-8f35-4128-ac4f-ba245f98fde9 service nova] Releasing lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.558689] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.558689] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.558689] env[61243]: DEBUG nova.network.neutron [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1134.805495] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.805753] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.806041] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.806228] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61243) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1134.807690] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55fed2dc-4b55-4094-8016-27d5186a7fa9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.820105] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f328ec7-0eae-4677-8dfd-4ff2ddaf967a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.841956] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c11bf99-068c-4601-8336-daa9ad45e86e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.850407] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ede7af91-9cf3-47ef-bf37-518bba6d3495 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.883825] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180584MB free_disk=154GB free_vcpus=48 pci_devices=None {{(pid=61243) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1134.884021] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.884202] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.089510] env[61243]: DEBUG nova.network.neutron [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Instance cache missing network info. {{(pid=61243) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1135.178567] env[61243]: DEBUG nova.compute.manager [req-6f32577d-f87d-4538-964e-f52d75e472d7 req-0c19b1ef-1f9b-413c-84ed-b469d01c8307 service nova] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Received event network-vif-plugged-2168e525-2705-4de2-9879-ac1a43eb9f84 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1135.178837] env[61243]: DEBUG oslo_concurrency.lockutils [req-6f32577d-f87d-4538-964e-f52d75e472d7 req-0c19b1ef-1f9b-413c-84ed-b469d01c8307 service nova] Acquiring lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.179020] env[61243]: DEBUG oslo_concurrency.lockutils [req-6f32577d-f87d-4538-964e-f52d75e472d7 req-0c19b1ef-1f9b-413c-84ed-b469d01c8307 service nova] Lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.179206] env[61243]: DEBUG oslo_concurrency.lockutils [req-6f32577d-f87d-4538-964e-f52d75e472d7 req-0c19b1ef-1f9b-413c-84ed-b469d01c8307 service nova] Lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.179368] env[61243]: DEBUG nova.compute.manager [req-6f32577d-f87d-4538-964e-f52d75e472d7 req-0c19b1ef-1f9b-413c-84ed-b469d01c8307 service nova] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] No waiting events found dispatching network-vif-plugged-2168e525-2705-4de2-9879-ac1a43eb9f84 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1135.179543] env[61243]: WARNING nova.compute.manager [req-6f32577d-f87d-4538-964e-f52d75e472d7 req-0c19b1ef-1f9b-413c-84ed-b469d01c8307 service nova] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Received unexpected event network-vif-plugged-2168e525-2705-4de2-9879-ac1a43eb9f84 for instance with vm_state building and task_state spawning. [ 1135.179720] env[61243]: DEBUG nova.compute.manager [req-6f32577d-f87d-4538-964e-f52d75e472d7 req-0c19b1ef-1f9b-413c-84ed-b469d01c8307 service nova] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Received event network-changed-2168e525-2705-4de2-9879-ac1a43eb9f84 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1135.179935] env[61243]: DEBUG nova.compute.manager [req-6f32577d-f87d-4538-964e-f52d75e472d7 req-0c19b1ef-1f9b-413c-84ed-b469d01c8307 service nova] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Refreshing instance network info cache due to event network-changed-2168e525-2705-4de2-9879-ac1a43eb9f84. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1135.180134] env[61243]: DEBUG oslo_concurrency.lockutils [req-6f32577d-f87d-4538-964e-f52d75e472d7 req-0c19b1ef-1f9b-413c-84ed-b469d01c8307 service nova] Acquiring lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.229324] env[61243]: DEBUG nova.network.neutron [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updating instance_info_cache with network_info: [{"id": "2168e525-2705-4de2-9879-ac1a43eb9f84", "address": "fa:16:3e:d0:d0:ab", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2168e525-27", "ovs_interfaceid": "2168e525-2705-4de2-9879-ac1a43eb9f84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.387597] env[61243]: DEBUG nova.compute.manager [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Start spawning the instance on the hypervisor. {{(pid=61243) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1135.388090] env[61243]: DEBUG nova.virt.hardware [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1135.388321] env[61243]: DEBUG nova.virt.hardware [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1135.388485] env[61243]: DEBUG nova.virt.hardware [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1135.388677] env[61243]: DEBUG nova.virt.hardware [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1135.388831] env[61243]: DEBUG nova.virt.hardware [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1135.388991] env[61243]: DEBUG nova.virt.hardware [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1135.389217] env[61243]: DEBUG nova.virt.hardware [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1135.389387] env[61243]: DEBUG nova.virt.hardware [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1135.389558] env[61243]: DEBUG nova.virt.hardware [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1135.389727] env[61243]: DEBUG nova.virt.hardware [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1135.389942] env[61243]: DEBUG nova.virt.hardware [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1135.393949] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83193737-c5f8-4997-b438-d6144240b05e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.402846] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2163ef3f-f04f-4cff-8364-8e3315a469c9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.732599] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.732816] env[61243]: DEBUG nova.compute.manager [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Instance network_info: |[{"id": "2168e525-2705-4de2-9879-ac1a43eb9f84", "address": "fa:16:3e:d0:d0:ab", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2168e525-27", "ovs_interfaceid": "2168e525-2705-4de2-9879-ac1a43eb9f84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61243) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1135.733206] env[61243]: DEBUG oslo_concurrency.lockutils [req-6f32577d-f87d-4538-964e-f52d75e472d7 req-0c19b1ef-1f9b-413c-84ed-b469d01c8307 service nova] Acquired lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.733437] env[61243]: DEBUG nova.network.neutron [req-6f32577d-f87d-4538-964e-f52d75e472d7 req-0c19b1ef-1f9b-413c-84ed-b469d01c8307 service nova] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Refreshing network info cache for port 2168e525-2705-4de2-9879-ac1a43eb9f84 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1135.734656] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:d0:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ff3ecd2f-0b10-4faf-a512-fd7a20c28df1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2168e525-2705-4de2-9879-ac1a43eb9f84', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1135.742489] env[61243]: DEBUG oslo.service.loopingcall [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1135.745386] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1135.745853] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-16802afb-4ecc-4721-8b62-65fbf9d41f2a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.767459] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1135.767459] env[61243]: value = "task-1339608" [ 1135.767459] env[61243]: _type = "Task" [ 1135.767459] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.778887] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339608, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.916039] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 3f839aa5-2b9a-4807-b63b-931f74455532 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1135.916216] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance d24f3683-169b-46a8-9387-b6fc88b654bd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1135.916417] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 91b14a22-811a-4151-9769-9c684ec993db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1135.916455] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 486ebf65-4376-424c-9deb-f1ef70e44be3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1135.916571] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 440b35e6-eecb-4e20-831d-cf0db0158e81 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1135.916687] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1135.916881] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1135.917061] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1135.954505] env[61243]: DEBUG nova.network.neutron [req-6f32577d-f87d-4538-964e-f52d75e472d7 req-0c19b1ef-1f9b-413c-84ed-b469d01c8307 service nova] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updated VIF entry in instance network info cache for port 2168e525-2705-4de2-9879-ac1a43eb9f84. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1135.954916] env[61243]: DEBUG nova.network.neutron [req-6f32577d-f87d-4538-964e-f52d75e472d7 req-0c19b1ef-1f9b-413c-84ed-b469d01c8307 service nova] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updating instance_info_cache with network_info: [{"id": "2168e525-2705-4de2-9879-ac1a43eb9f84", "address": "fa:16:3e:d0:d0:ab", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2168e525-27", "ovs_interfaceid": "2168e525-2705-4de2-9879-ac1a43eb9f84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.005513] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8884f421-bb79-415a-b84c-f69ce6529f66 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.013666] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a183d5-c3cd-4822-99bb-9b1b612815c5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.045153] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc0ee23-9959-4059-81c9-9531c66d71dc {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.052834] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-415cfefd-b28c-422f-bd8c-35f1a70c2a9b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.067847] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1136.278263] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339608, 'name': CreateVM_Task, 'duration_secs': 0.358649} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.278482] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1136.279124] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'guest_format': None, 'device_type': None, 'boot_index': 0, 'mount_device': '/dev/sda', 'attachment_id': '13fe082d-ccef-455b-b56e-09f416feca55', 'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285805', 'volume_id': 'd7581ea4-bb02-421e-a48c-c34488368a12', 'name': 'volume-d7581ea4-bb02-421e-a48c-c34488368a12', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4', 'attached_at': '', 'detached_at': '', 'volume_id': 'd7581ea4-bb02-421e-a48c-c34488368a12', 'serial': 'd7581ea4-bb02-421e-a48c-c34488368a12'}, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=61243) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1136.279355] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Root volume attach. Driver type: vmdk {{(pid=61243) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1136.280112] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da74b8bd-057e-47fd-9dba-5203fdf0fb8a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.287229] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96bb9466-b0bf-4539-b8db-e733f7044ee5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.292870] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b6ea40c-f18d-4e0c-a48e-d2e5eafad5d3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.298231] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-26d12969-66ff-4c93-851f-95f05ec6eb1f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.305360] env[61243]: DEBUG oslo_vmware.api [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1136.305360] env[61243]: value = "task-1339609" [ 1136.305360] env[61243]: _type = "Task" [ 1136.305360] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.312444] env[61243]: DEBUG oslo_vmware.api [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339609, 'name': RelocateVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.460098] env[61243]: DEBUG oslo_concurrency.lockutils [req-6f32577d-f87d-4538-964e-f52d75e472d7 req-0c19b1ef-1f9b-413c-84ed-b469d01c8307 service nova] Releasing lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.571217] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1136.815490] env[61243]: DEBUG oslo_vmware.api [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339609, 'name': RelocateVM_Task, 'duration_secs': 0.34366} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.815737] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Volume attach. Driver type: vmdk {{(pid=61243) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1136.815946] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285805', 'volume_id': 'd7581ea4-bb02-421e-a48c-c34488368a12', 'name': 'volume-d7581ea4-bb02-421e-a48c-c34488368a12', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4', 'attached_at': '', 'detached_at': '', 'volume_id': 'd7581ea4-bb02-421e-a48c-c34488368a12', 'serial': 'd7581ea4-bb02-421e-a48c-c34488368a12'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1136.816736] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-071213bf-593f-4583-9ea5-5d4c5f14851a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.832064] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0bcc383-12e3-4277-93cd-5534444d18e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.854723] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] volume-d7581ea4-bb02-421e-a48c-c34488368a12/volume-d7581ea4-bb02-421e-a48c-c34488368a12.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1136.854994] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4e726b2-c3d3-490a-b932-3b51f3f921a2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.874276] env[61243]: DEBUG oslo_vmware.api [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1136.874276] env[61243]: value = "task-1339610" [ 1136.874276] env[61243]: _type = "Task" [ 1136.874276] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.882111] env[61243]: DEBUG oslo_vmware.api [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339610, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.076481] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61243) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1137.076648] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.192s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.386203] env[61243]: DEBUG oslo_vmware.api [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339610, 'name': ReconfigVM_Task, 'duration_secs': 0.283674} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.386514] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Reconfigured VM instance instance-00000071 to attach disk [datastore2] volume-d7581ea4-bb02-421e-a48c-c34488368a12/volume-d7581ea4-bb02-421e-a48c-c34488368a12.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1137.391608] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3cdcefb0-1ffc-4719-ab7a-9e526dc91390 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.409068] env[61243]: DEBUG oslo_vmware.api [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1137.409068] env[61243]: value = "task-1339611" [ 1137.409068] env[61243]: _type = "Task" [ 1137.409068] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.418210] env[61243]: DEBUG oslo_vmware.api [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339611, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.918764] env[61243]: DEBUG oslo_vmware.api [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339611, 'name': ReconfigVM_Task, 'duration_secs': 0.134361} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.919107] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285805', 'volume_id': 'd7581ea4-bb02-421e-a48c-c34488368a12', 'name': 'volume-d7581ea4-bb02-421e-a48c-c34488368a12', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4', 'attached_at': '', 'detached_at': '', 'volume_id': 'd7581ea4-bb02-421e-a48c-c34488368a12', 'serial': 'd7581ea4-bb02-421e-a48c-c34488368a12'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1137.919581] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d156cc12-7b69-43f9-8dd7-b03826805bd2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.925936] env[61243]: DEBUG oslo_vmware.api [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1137.925936] env[61243]: value = "task-1339612" [ 1137.925936] env[61243]: _type = "Task" [ 1137.925936] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.934198] env[61243]: DEBUG oslo_vmware.api [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339612, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.077023] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.077289] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.077464] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.302804] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.435403] env[61243]: DEBUG oslo_vmware.api [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339612, 'name': Rename_Task, 'duration_secs': 0.463022} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.435686] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1138.435941] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-050e066b-1149-40ea-919f-66d7181d31c4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.443680] env[61243]: DEBUG oslo_vmware.api [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1138.443680] env[61243]: value = "task-1339613" [ 1138.443680] env[61243]: _type = "Task" [ 1138.443680] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.453561] env[61243]: DEBUG oslo_vmware.api [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339613, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.954193] env[61243]: DEBUG oslo_vmware.api [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339613, 'name': PowerOnVM_Task, 'duration_secs': 0.474717} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.954600] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1138.954735] env[61243]: INFO nova.compute.manager [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Took 3.57 seconds to spawn the instance on the hypervisor. [ 1138.954875] env[61243]: DEBUG nova.compute.manager [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1138.955651] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103a28f3-3b86-4ee9-8029-e3cda1980260 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.302320] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1139.302618] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Starting heal instance info cache {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1139.474808] env[61243]: INFO nova.compute.manager [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Took 10.05 seconds to build instance. [ 1139.977029] env[61243]: DEBUG oslo_concurrency.lockutils [None req-8aec93d2-804a-425e-a946-02c762b4dfbf tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.563s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.334964] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.335162] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquired lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.335319] env[61243]: DEBUG nova.network.neutron [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Forcefully refreshing network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1140.586849] env[61243]: DEBUG nova.compute.manager [req-8029a22d-14f2-4af9-a0f1-da00daefb7f7 req-062fb439-0ef0-4b52-b153-55070a6bb2b7 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Received event network-changed-a0b324d6-6a73-4956-85c3-38f755003b74 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1140.586972] env[61243]: DEBUG nova.compute.manager [req-8029a22d-14f2-4af9-a0f1-da00daefb7f7 req-062fb439-0ef0-4b52-b153-55070a6bb2b7 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Refreshing instance network info cache due to event network-changed-a0b324d6-6a73-4956-85c3-38f755003b74. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1140.587460] env[61243]: DEBUG oslo_concurrency.lockutils [req-8029a22d-14f2-4af9-a0f1-da00daefb7f7 req-062fb439-0ef0-4b52-b153-55070a6bb2b7 service nova] Acquiring lock "refresh_cache-3f839aa5-2b9a-4807-b63b-931f74455532" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.587701] env[61243]: DEBUG oslo_concurrency.lockutils [req-8029a22d-14f2-4af9-a0f1-da00daefb7f7 req-062fb439-0ef0-4b52-b153-55070a6bb2b7 service nova] Acquired lock "refresh_cache-3f839aa5-2b9a-4807-b63b-931f74455532" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.587784] env[61243]: DEBUG nova.network.neutron [req-8029a22d-14f2-4af9-a0f1-da00daefb7f7 req-062fb439-0ef0-4b52-b153-55070a6bb2b7 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Refreshing network info cache for port a0b324d6-6a73-4956-85c3-38f755003b74 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1141.275923] env[61243]: DEBUG nova.network.neutron [req-8029a22d-14f2-4af9-a0f1-da00daefb7f7 req-062fb439-0ef0-4b52-b153-55070a6bb2b7 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Updated VIF entry in instance network info cache for port a0b324d6-6a73-4956-85c3-38f755003b74. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1141.276323] env[61243]: DEBUG nova.network.neutron [req-8029a22d-14f2-4af9-a0f1-da00daefb7f7 req-062fb439-0ef0-4b52-b153-55070a6bb2b7 service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Updating instance_info_cache with network_info: [{"id": "a0b324d6-6a73-4956-85c3-38f755003b74", "address": "fa:16:3e:16:b3:46", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0b324d6-6a", "ovs_interfaceid": "a0b324d6-6a73-4956-85c3-38f755003b74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.562923] env[61243]: DEBUG nova.network.neutron [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Updating instance_info_cache with network_info: [{"id": "00f9d661-7722-4165-8933-c025ac1f938a", "address": "fa:16:3e:e9:b9:d7", "network": {"id": "62b19b92-0395-4f58-9071-61e067e1e903", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-41471245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2465299214be4d5f877f340489215d6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00f9d661-77", "ovs_interfaceid": "00f9d661-7722-4165-8933-c025ac1f938a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.778754] env[61243]: DEBUG oslo_concurrency.lockutils [req-8029a22d-14f2-4af9-a0f1-da00daefb7f7 req-062fb439-0ef0-4b52-b153-55070a6bb2b7 service nova] Releasing lock "refresh_cache-3f839aa5-2b9a-4807-b63b-931f74455532" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.038500] env[61243]: DEBUG nova.compute.manager [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Stashing vm_state: active {{(pid=61243) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1142.065599] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Releasing lock "refresh_cache-486ebf65-4376-424c-9deb-f1ef70e44be3" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.065790] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Updated the network info_cache for instance {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1142.066046] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.066188] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61243) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1142.558263] env[61243]: DEBUG oslo_concurrency.lockutils [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.558537] env[61243]: DEBUG oslo_concurrency.lockutils [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.611191] env[61243]: DEBUG nova.compute.manager [req-20a72b47-ca05-4904-91cb-9e2723c90454 req-d4f0d1e9-fd27-42c2-b0fb-612a8c5cc81e service nova] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Received event network-changed-2168e525-2705-4de2-9879-ac1a43eb9f84 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1142.611407] env[61243]: DEBUG nova.compute.manager [req-20a72b47-ca05-4904-91cb-9e2723c90454 req-d4f0d1e9-fd27-42c2-b0fb-612a8c5cc81e service nova] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Refreshing instance network info cache due to event network-changed-2168e525-2705-4de2-9879-ac1a43eb9f84. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1142.611624] env[61243]: DEBUG oslo_concurrency.lockutils [req-20a72b47-ca05-4904-91cb-9e2723c90454 req-d4f0d1e9-fd27-42c2-b0fb-612a8c5cc81e service nova] Acquiring lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.611919] env[61243]: DEBUG oslo_concurrency.lockutils [req-20a72b47-ca05-4904-91cb-9e2723c90454 req-d4f0d1e9-fd27-42c2-b0fb-612a8c5cc81e service nova] Acquired lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.612191] env[61243]: DEBUG nova.network.neutron [req-20a72b47-ca05-4904-91cb-9e2723c90454 req-d4f0d1e9-fd27-42c2-b0fb-612a8c5cc81e service nova] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Refreshing network info cache for port 2168e525-2705-4de2-9879-ac1a43eb9f84 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1143.063019] env[61243]: INFO nova.compute.claims [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1143.389054] env[61243]: DEBUG nova.network.neutron [req-20a72b47-ca05-4904-91cb-9e2723c90454 req-d4f0d1e9-fd27-42c2-b0fb-612a8c5cc81e service nova] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updated VIF entry in instance network info cache for port 2168e525-2705-4de2-9879-ac1a43eb9f84. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1143.389054] env[61243]: DEBUG nova.network.neutron [req-20a72b47-ca05-4904-91cb-9e2723c90454 req-d4f0d1e9-fd27-42c2-b0fb-612a8c5cc81e service nova] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updating instance_info_cache with network_info: [{"id": "2168e525-2705-4de2-9879-ac1a43eb9f84", "address": "fa:16:3e:d0:d0:ab", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2168e525-27", "ovs_interfaceid": "2168e525-2705-4de2-9879-ac1a43eb9f84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.411417] env[61243]: DEBUG oslo_concurrency.lockutils [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "d24f3683-169b-46a8-9387-b6fc88b654bd" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.411647] env[61243]: DEBUG oslo_concurrency.lockutils [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d24f3683-169b-46a8-9387-b6fc88b654bd" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.569812] env[61243]: INFO nova.compute.resource_tracker [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updating resource usage from migration bc1e3cd7-8657-4772-903c-ca1056270095 [ 1143.651014] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a34d0053-011e-4c6d-91b6-b82569ca7935 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.658932] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887836ad-1259-46a3-8e43-b0c00906c53c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.688251] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ada74b9-bb9e-43f8-9e34-0bef61785384 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.694964] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1249530c-a30a-4022-a965-d7d9a028e3a2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.707861] env[61243]: DEBUG nova.compute.provider_tree [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.890766] env[61243]: DEBUG oslo_concurrency.lockutils [req-20a72b47-ca05-4904-91cb-9e2723c90454 req-d4f0d1e9-fd27-42c2-b0fb-612a8c5cc81e service nova] Releasing lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.915064] env[61243]: DEBUG nova.compute.utils [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Using /dev/sd instead of None {{(pid=61243) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1144.211348] env[61243]: DEBUG nova.scheduler.client.report [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1144.418340] env[61243]: DEBUG oslo_concurrency.lockutils [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d24f3683-169b-46a8-9387-b6fc88b654bd" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.716425] env[61243]: DEBUG oslo_concurrency.lockutils [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.158s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.716763] env[61243]: INFO nova.compute.manager [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Migrating [ 1145.231987] env[61243]: DEBUG oslo_concurrency.lockutils [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1145.232223] env[61243]: DEBUG oslo_concurrency.lockutils [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.232409] env[61243]: DEBUG nova.network.neutron [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1145.478210] env[61243]: DEBUG oslo_concurrency.lockutils [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "d24f3683-169b-46a8-9387-b6fc88b654bd" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.478616] env[61243]: DEBUG oslo_concurrency.lockutils [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d24f3683-169b-46a8-9387-b6fc88b654bd" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.479029] env[61243]: INFO nova.compute.manager [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Attaching volume 33f3229d-c0fb-4829-950d-6b9397e19d64 to /dev/sdb [ 1145.511761] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659fe72a-35dd-4b46-adaa-b139bb9f6204 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.519122] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62cf477-3461-4d0c-b2da-ae59a6e5e0b6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.532573] env[61243]: DEBUG nova.virt.block_device [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Updating existing volume attachment record: f6d125d4-6c78-4914-a98b-f27d0b176096 {{(pid=61243) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1145.959834] env[61243]: DEBUG nova.network.neutron [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updating instance_info_cache with network_info: [{"id": "2168e525-2705-4de2-9879-ac1a43eb9f84", "address": "fa:16:3e:d0:d0:ab", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2168e525-27", "ovs_interfaceid": "2168e525-2705-4de2-9879-ac1a43eb9f84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.462833] env[61243]: DEBUG oslo_concurrency.lockutils [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.979121] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d0c67aa-4bc4-4689-969f-025f8214f28d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.996888] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updating instance '9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4' progress to 0 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1148.504020] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1148.504376] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f3c73d97-0779-43ae-a380-0eaa222fb476 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.511808] env[61243]: DEBUG oslo_vmware.api [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1148.511808] env[61243]: value = "task-1339616" [ 1148.511808] env[61243]: _type = "Task" [ 1148.511808] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.520086] env[61243]: DEBUG oslo_vmware.api [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339616, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.023115] env[61243]: DEBUG oslo_vmware.api [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339616, 'name': PowerOffVM_Task, 'duration_secs': 0.191521} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.023423] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1149.023607] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updating instance '9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4' progress to 17 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1149.529374] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1149.529630] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1149.529812] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1149.530023] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1149.530179] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1149.530332] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1149.530542] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1149.530709] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1149.530881] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1149.531061] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1149.531245] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1149.536179] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2a9faa7-4c69-4171-94b4-9bb0afbf7256 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.551336] env[61243]: DEBUG oslo_vmware.api [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1149.551336] env[61243]: value = "task-1339617" [ 1149.551336] env[61243]: _type = "Task" [ 1149.551336] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.559215] env[61243]: DEBUG oslo_vmware.api [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339617, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.061044] env[61243]: DEBUG oslo_vmware.api [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339617, 'name': ReconfigVM_Task, 'duration_secs': 0.169648} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.061437] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updating instance '9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4' progress to 33 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1150.075046] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Volume attach. Driver type: vmdk {{(pid=61243) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1150.075287] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285808', 'volume_id': '33f3229d-c0fb-4829-950d-6b9397e19d64', 'name': 'volume-33f3229d-c0fb-4829-950d-6b9397e19d64', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd24f3683-169b-46a8-9387-b6fc88b654bd', 'attached_at': '', 'detached_at': '', 'volume_id': '33f3229d-c0fb-4829-950d-6b9397e19d64', 'serial': '33f3229d-c0fb-4829-950d-6b9397e19d64'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1150.076082] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d1d904-2eca-4534-8d1f-807fbf3ccec4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.091482] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26801cc-cdc4-4401-becf-91cb07d412c7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.115439] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] volume-33f3229d-c0fb-4829-950d-6b9397e19d64/volume-33f3229d-c0fb-4829-950d-6b9397e19d64.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1150.116382] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a45cbb3-7442-4472-b951-655a6888e70b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.134705] env[61243]: DEBUG oslo_vmware.api [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1150.134705] env[61243]: value = "task-1339618" [ 1150.134705] env[61243]: _type = "Task" [ 1150.134705] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.142637] env[61243]: DEBUG oslo_vmware.api [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339618, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.567810] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1150.568093] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1150.568253] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1150.568444] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1150.568608] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1150.568768] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1150.568985] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1150.569169] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1150.569345] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1150.569515] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1150.569734] env[61243]: DEBUG nova.virt.hardware [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1150.574983] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Reconfiguring VM instance instance-00000071 to detach disk 2000 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1150.575286] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2a15f92-bb41-489d-b483-08910962b4b7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.593900] env[61243]: DEBUG oslo_vmware.api [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1150.593900] env[61243]: value = "task-1339619" [ 1150.593900] env[61243]: _type = "Task" [ 1150.593900] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.601831] env[61243]: DEBUG oslo_vmware.api [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339619, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.642999] env[61243]: DEBUG oslo_vmware.api [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339618, 'name': ReconfigVM_Task, 'duration_secs': 0.346871} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.643291] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Reconfigured VM instance instance-0000006e to attach disk [datastore2] volume-33f3229d-c0fb-4829-950d-6b9397e19d64/volume-33f3229d-c0fb-4829-950d-6b9397e19d64.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1150.648034] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c355959-114b-44de-8856-6f587cdfb97c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.662310] env[61243]: DEBUG oslo_vmware.api [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1150.662310] env[61243]: value = "task-1339620" [ 1150.662310] env[61243]: _type = "Task" [ 1150.662310] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.669927] env[61243]: DEBUG oslo_vmware.api [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339620, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.104039] env[61243]: DEBUG oslo_vmware.api [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339619, 'name': ReconfigVM_Task, 'duration_secs': 0.173631} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.104375] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Reconfigured VM instance instance-00000071 to detach disk 2000 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1151.105097] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109fcbac-94e0-4318-b1aa-97cc8581370b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.127344] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] volume-d7581ea4-bb02-421e-a48c-c34488368a12/volume-d7581ea4-bb02-421e-a48c-c34488368a12.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1151.127618] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca2ecf06-41a1-48aa-aad3-d3702228c197 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.145661] env[61243]: DEBUG oslo_vmware.api [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1151.145661] env[61243]: value = "task-1339621" [ 1151.145661] env[61243]: _type = "Task" [ 1151.145661] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.153437] env[61243]: DEBUG oslo_vmware.api [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339621, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.172298] env[61243]: DEBUG oslo_vmware.api [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339620, 'name': ReconfigVM_Task, 'duration_secs': 0.176243} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.172407] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285808', 'volume_id': '33f3229d-c0fb-4829-950d-6b9397e19d64', 'name': 'volume-33f3229d-c0fb-4829-950d-6b9397e19d64', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd24f3683-169b-46a8-9387-b6fc88b654bd', 'attached_at': '', 'detached_at': '', 'volume_id': '33f3229d-c0fb-4829-950d-6b9397e19d64', 'serial': '33f3229d-c0fb-4829-950d-6b9397e19d64'} {{(pid=61243) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1151.656434] env[61243]: DEBUG oslo_vmware.api [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339621, 'name': ReconfigVM_Task, 'duration_secs': 0.247709} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.656742] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Reconfigured VM instance instance-00000071 to attach disk [datastore2] volume-d7581ea4-bb02-421e-a48c-c34488368a12/volume-d7581ea4-bb02-421e-a48c-c34488368a12.vmdk or device None with type thin {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1151.657030] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updating instance '9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4' progress to 50 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1152.164592] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f91d02-a5e9-4dfa-8877-760955aa4b3a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.183317] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab0f33c-9507-4422-96e0-9eea823f9e07 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.200436] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updating instance '9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4' progress to 67 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1152.208909] env[61243]: DEBUG nova.objects.instance [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lazy-loading 'flavor' on Instance uuid d24f3683-169b-46a8-9387-b6fc88b654bd {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1152.714728] env[61243]: DEBUG oslo_concurrency.lockutils [None req-470c9944-99bf-461b-90b5-89d14b9eaaa6 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d24f3683-169b-46a8-9387-b6fc88b654bd" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.236s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.887327] env[61243]: DEBUG oslo_concurrency.lockutils [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "d24f3683-169b-46a8-9387-b6fc88b654bd" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.887581] env[61243]: DEBUG oslo_concurrency.lockutils [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d24f3683-169b-46a8-9387-b6fc88b654bd" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.390635] env[61243]: INFO nova.compute.manager [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Detaching volume 33f3229d-c0fb-4829-950d-6b9397e19d64 [ 1153.421671] env[61243]: INFO nova.virt.block_device [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Attempting to driver detach volume 33f3229d-c0fb-4829-950d-6b9397e19d64 from mountpoint /dev/sdb [ 1153.421936] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Volume detach. Driver type: vmdk {{(pid=61243) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1153.422152] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285808', 'volume_id': '33f3229d-c0fb-4829-950d-6b9397e19d64', 'name': 'volume-33f3229d-c0fb-4829-950d-6b9397e19d64', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd24f3683-169b-46a8-9387-b6fc88b654bd', 'attached_at': '', 'detached_at': '', 'volume_id': '33f3229d-c0fb-4829-950d-6b9397e19d64', 'serial': '33f3229d-c0fb-4829-950d-6b9397e19d64'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1153.423119] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8934206-bbfb-403a-a966-4e2f0b908282 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.445364] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e38834-ab3f-452e-b10e-6d02571fa573 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.452217] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fcbc619-4c92-421b-9ce1-308d0f90773b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.471331] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9848120b-2ec9-4b3e-8f9c-168cf3dcd299 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.485106] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] The volume has not been displaced from its original location: [datastore2] volume-33f3229d-c0fb-4829-950d-6b9397e19d64/volume-33f3229d-c0fb-4829-950d-6b9397e19d64.vmdk. No consolidation needed. {{(pid=61243) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1153.490198] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Reconfiguring VM instance instance-0000006e to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1153.490464] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ec1cf96-9354-4348-ab08-a5e0f4d198d4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.506858] env[61243]: DEBUG oslo_vmware.api [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1153.506858] env[61243]: value = "task-1339622" [ 1153.506858] env[61243]: _type = "Task" [ 1153.506858] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.514098] env[61243]: DEBUG oslo_vmware.api [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339622, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.834414] env[61243]: DEBUG nova.network.neutron [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Port 2168e525-2705-4de2-9879-ac1a43eb9f84 binding to destination host cpu-1 is already ACTIVE {{(pid=61243) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1154.016622] env[61243]: DEBUG oslo_vmware.api [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339622, 'name': ReconfigVM_Task, 'duration_secs': 0.205304} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.016981] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Reconfigured VM instance instance-0000006e to detach disk 2001 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1154.021485] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce98158c-05f6-4c86-90b7-10339f667e9b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.035877] env[61243]: DEBUG oslo_vmware.api [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1154.035877] env[61243]: value = "task-1339623" [ 1154.035877] env[61243]: _type = "Task" [ 1154.035877] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.043694] env[61243]: DEBUG oslo_vmware.api [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339623, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.545400] env[61243]: DEBUG oslo_vmware.api [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339623, 'name': ReconfigVM_Task, 'duration_secs': 0.132644} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.545724] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285808', 'volume_id': '33f3229d-c0fb-4829-950d-6b9397e19d64', 'name': 'volume-33f3229d-c0fb-4829-950d-6b9397e19d64', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd24f3683-169b-46a8-9387-b6fc88b654bd', 'attached_at': '', 'detached_at': '', 'volume_id': '33f3229d-c0fb-4829-950d-6b9397e19d64', 'serial': '33f3229d-c0fb-4829-950d-6b9397e19d64'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1154.855823] env[61243]: DEBUG oslo_concurrency.lockutils [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.856033] env[61243]: DEBUG oslo_concurrency.lockutils [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.856221] env[61243]: DEBUG oslo_concurrency.lockutils [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.086766] env[61243]: DEBUG nova.objects.instance [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lazy-loading 'flavor' on Instance uuid d24f3683-169b-46a8-9387-b6fc88b654bd {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1155.890934] env[61243]: DEBUG oslo_concurrency.lockutils [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.891215] env[61243]: DEBUG oslo_concurrency.lockutils [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.891325] env[61243]: DEBUG nova.network.neutron [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1156.093993] env[61243]: DEBUG oslo_concurrency.lockutils [None req-db41c4c3-2211-40e7-b0d8-ed699a395aa2 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d24f3683-169b-46a8-9387-b6fc88b654bd" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.206s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.593213] env[61243]: DEBUG nova.network.neutron [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updating instance_info_cache with network_info: [{"id": "2168e525-2705-4de2-9879-ac1a43eb9f84", "address": "fa:16:3e:d0:d0:ab", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2168e525-27", "ovs_interfaceid": "2168e525-2705-4de2-9879-ac1a43eb9f84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.096463] env[61243]: DEBUG oslo_concurrency.lockutils [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.118927] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "d24f3683-169b-46a8-9387-b6fc88b654bd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.119207] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d24f3683-169b-46a8-9387-b6fc88b654bd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.119423] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "d24f3683-169b-46a8-9387-b6fc88b654bd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.119614] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d24f3683-169b-46a8-9387-b6fc88b654bd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.119787] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d24f3683-169b-46a8-9387-b6fc88b654bd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.121806] env[61243]: INFO nova.compute.manager [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Terminating instance [ 1157.124266] env[61243]: DEBUG nova.compute.manager [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1157.124266] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1157.124610] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df9e5c9f-23a2-4c98-a4a6-10e22f3b8d90 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.132566] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1157.132800] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8bec1f19-0025-4598-9257-216b41fac831 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.139804] env[61243]: DEBUG oslo_vmware.api [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1157.139804] env[61243]: value = "task-1339624" [ 1157.139804] env[61243]: _type = "Task" [ 1157.139804] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.150551] env[61243]: DEBUG oslo_vmware.api [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339624, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.607556] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82031792-e146-4bc3-abf8-7936923d1da0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.617972] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0192b4d9-d750-4fc4-bb23-951332bbc991 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.653467] env[61243]: DEBUG oslo_vmware.api [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339624, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.151514] env[61243]: DEBUG oslo_vmware.api [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339624, 'name': PowerOffVM_Task, 'duration_secs': 0.834085} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.151897] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1158.151987] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1158.152248] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6d47d298-06ce-4470-926c-ec413dc73249 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.225559] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1158.225793] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1158.225980] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Deleting the datastore file [datastore2] d24f3683-169b-46a8-9387-b6fc88b654bd {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1158.226276] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cfbbc5d0-085f-40b6-959b-3897cc7369c0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.232796] env[61243]: DEBUG oslo_vmware.api [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for the task: (returnval){ [ 1158.232796] env[61243]: value = "task-1339626" [ 1158.232796] env[61243]: _type = "Task" [ 1158.232796] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.239938] env[61243]: DEBUG oslo_vmware.api [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339626, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.713995] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d10d5a23-e8c8-485f-bffc-e631f2d16672 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.732221] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e229000-80bf-41e3-9e4e-af9f335915ef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.741491] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updating instance '9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4' progress to 83 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1158.747494] env[61243]: DEBUG oslo_vmware.api [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Task: {'id': task-1339626, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128617} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.747935] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1158.748140] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1158.748325] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1158.748504] env[61243]: INFO nova.compute.manager [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Took 1.62 seconds to destroy the instance on the hypervisor. [ 1158.748744] env[61243]: DEBUG oslo.service.loopingcall [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1158.748936] env[61243]: DEBUG nova.compute.manager [-] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1158.749049] env[61243]: DEBUG nova.network.neutron [-] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1159.194333] env[61243]: DEBUG nova.compute.manager [req-e9e484de-0a5b-46e6-a811-70a3e15de4d7 req-51f63c56-a706-4472-929d-7c79963af781 service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Received event network-vif-deleted-1c517f1d-301b-43a0-8ae9-6099f75580ed {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1159.194569] env[61243]: INFO nova.compute.manager [req-e9e484de-0a5b-46e6-a811-70a3e15de4d7 req-51f63c56-a706-4472-929d-7c79963af781 service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Neutron deleted interface 1c517f1d-301b-43a0-8ae9-6099f75580ed; detaching it from the instance and deleting it from the info cache [ 1159.194701] env[61243]: DEBUG nova.network.neutron [req-e9e484de-0a5b-46e6-a811-70a3e15de4d7 req-51f63c56-a706-4472-929d-7c79963af781 service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.250365] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1159.251386] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f220216d-0b65-40bf-b5d3-95ceef2fb270 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.258894] env[61243]: DEBUG oslo_vmware.api [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1159.258894] env[61243]: value = "task-1339627" [ 1159.258894] env[61243]: _type = "Task" [ 1159.258894] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.268545] env[61243]: DEBUG oslo_vmware.api [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339627, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.673037] env[61243]: DEBUG nova.network.neutron [-] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.697057] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca5ef106-9efe-441a-a165-5f5e5b72fffb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.707261] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b24845-2985-40d4-910e-985cc7e53778 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.737402] env[61243]: DEBUG nova.compute.manager [req-e9e484de-0a5b-46e6-a811-70a3e15de4d7 req-51f63c56-a706-4472-929d-7c79963af781 service nova] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Detach interface failed, port_id=1c517f1d-301b-43a0-8ae9-6099f75580ed, reason: Instance d24f3683-169b-46a8-9387-b6fc88b654bd could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1159.767848] env[61243]: DEBUG oslo_vmware.api [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339627, 'name': PowerOnVM_Task, 'duration_secs': 0.46388} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.768910] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1159.768910] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-15fc3e3f-d06f-4e8b-85b2-2f4fe6754ef3 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updating instance '9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4' progress to 100 {{(pid=61243) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1160.175469] env[61243]: INFO nova.compute.manager [-] [instance: d24f3683-169b-46a8-9387-b6fc88b654bd] Took 1.43 seconds to deallocate network for instance. [ 1160.682391] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.682772] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.682829] env[61243]: DEBUG nova.objects.instance [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lazy-loading 'resources' on Instance uuid d24f3683-169b-46a8-9387-b6fc88b654bd {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1161.285154] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc89dae5-9900-4b1c-a769-66b79458e73d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.293534] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb66a97b-7a5a-415b-bbf4-fac279a94875 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.331341] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3afdd3e8-c0b1-4fbd-a4d4-0a1ac94385ff {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.340422] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ceb2662-6ba9-4abe-8510-f969d0ac95fe {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.355818] env[61243]: DEBUG nova.compute.provider_tree [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1161.767075] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.767075] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.767075] env[61243]: DEBUG nova.compute.manager [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Going to confirm migration 6 {{(pid=61243) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1161.856715] env[61243]: DEBUG nova.scheduler.client.report [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1162.350191] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1162.350399] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquired lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.350585] env[61243]: DEBUG nova.network.neutron [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1162.350778] env[61243]: DEBUG nova.objects.instance [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lazy-loading 'info_cache' on Instance uuid 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1162.361208] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.678s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.384990] env[61243]: INFO nova.scheduler.client.report [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Deleted allocations for instance d24f3683-169b-46a8-9387-b6fc88b654bd [ 1162.891704] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b343a21e-bfa4-4e4d-868c-31df5ad93120 tempest-AttachVolumeNegativeTest-1386034266 tempest-AttachVolumeNegativeTest-1386034266-project-member] Lock "d24f3683-169b-46a8-9387-b6fc88b654bd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.772s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.564176] env[61243]: DEBUG nova.network.neutron [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updating instance_info_cache with network_info: [{"id": "2168e525-2705-4de2-9879-ac1a43eb9f84", "address": "fa:16:3e:d0:d0:ab", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2168e525-27", "ovs_interfaceid": "2168e525-2705-4de2-9879-ac1a43eb9f84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.067249] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Releasing lock "refresh_cache-9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1164.067592] env[61243]: DEBUG nova.objects.instance [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lazy-loading 'migration_context' on Instance uuid 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1164.570413] env[61243]: DEBUG nova.objects.base [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Object Instance<9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4> lazy-loaded attributes: info_cache,migration_context {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1164.571355] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec5dc97-6fe1-46e7-aa92-41c963844a3d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.591172] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b622d99b-6bcf-4240-891f-7e7bb7b29d06 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.596438] env[61243]: DEBUG oslo_vmware.api [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1164.596438] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52432040-a0d0-50be-8119-3da2050a53d8" [ 1164.596438] env[61243]: _type = "Task" [ 1164.596438] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.604146] env[61243]: DEBUG oslo_vmware.api [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52432040-a0d0-50be-8119-3da2050a53d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.107116] env[61243]: DEBUG oslo_vmware.api [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52432040-a0d0-50be-8119-3da2050a53d8, 'name': SearchDatastore_Task, 'duration_secs': 0.008097} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.107460] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.107626] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.353756] env[61243]: DEBUG oslo_concurrency.lockutils [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "486ebf65-4376-424c-9deb-f1ef70e44be3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.354163] env[61243]: DEBUG oslo_concurrency.lockutils [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.354484] env[61243]: DEBUG oslo_concurrency.lockutils [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "486ebf65-4376-424c-9deb-f1ef70e44be3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.354748] env[61243]: DEBUG oslo_concurrency.lockutils [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.355040] env[61243]: DEBUG oslo_concurrency.lockutils [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.357787] env[61243]: INFO nova.compute.manager [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Terminating instance [ 1165.359961] env[61243]: DEBUG nova.compute.manager [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1165.360259] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1165.361203] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b57246a-6fa1-43d0-a8b8-fc0efa9470f1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.370681] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1165.371054] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70ba4ce5-fc14-4f16-9964-b5adad6cfbc4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.378639] env[61243]: DEBUG oslo_vmware.api [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1165.378639] env[61243]: value = "task-1339629" [ 1165.378639] env[61243]: _type = "Task" [ 1165.378639] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.387016] env[61243]: DEBUG oslo_vmware.api [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339629, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.768981] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e374b50f-aabb-4795-ab24-cada3b31f65d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.776680] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dcb70c4-2ef4-42c4-b8ca-0bde34a2fbce {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.806893] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c4a28fa-665e-40fd-a4a9-e7e6917b59b2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.814119] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8ca6d52-df5d-4ac4-9ae6-a6d7af39d5ef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.826859] env[61243]: DEBUG nova.compute.provider_tree [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1165.888268] env[61243]: DEBUG oslo_vmware.api [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339629, 'name': PowerOffVM_Task, 'duration_secs': 0.251658} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.888475] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1165.888655] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1165.888897] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9099f3e0-7ba5-4387-a2b5-8609e0a55bf0 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.329984] env[61243]: DEBUG nova.scheduler.client.report [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1166.568221] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1166.568531] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Deleting contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1166.568804] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Deleting the datastore file [datastore1] 486ebf65-4376-424c-9deb-f1ef70e44be3 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1166.569138] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-564dd8ed-01bc-4a4a-bcc8-cc38039d21ad {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.576105] env[61243]: DEBUG oslo_vmware.api [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for the task: (returnval){ [ 1166.576105] env[61243]: value = "task-1339631" [ 1166.576105] env[61243]: _type = "Task" [ 1166.576105] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.583979] env[61243]: DEBUG oslo_vmware.api [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339631, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.085885] env[61243]: DEBUG oslo_vmware.api [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Task: {'id': task-1339631, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150202} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.086247] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1167.086448] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Deleted contents of the VM from datastore datastore1 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1167.086629] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1167.086813] env[61243]: INFO nova.compute.manager [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Took 1.73 seconds to destroy the instance on the hypervisor. [ 1167.087087] env[61243]: DEBUG oslo.service.loopingcall [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1167.087279] env[61243]: DEBUG nova.compute.manager [-] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1167.087375] env[61243]: DEBUG nova.network.neutron [-] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1167.340575] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.233s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.508526] env[61243]: DEBUG nova.compute.manager [req-e1b5bd54-9ea3-45af-b531-e07d25477ee3 req-3a9c1aad-f757-4469-aa98-804407a82e54 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Received event network-vif-deleted-00f9d661-7722-4165-8933-c025ac1f938a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1167.508576] env[61243]: INFO nova.compute.manager [req-e1b5bd54-9ea3-45af-b531-e07d25477ee3 req-3a9c1aad-f757-4469-aa98-804407a82e54 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Neutron deleted interface 00f9d661-7722-4165-8933-c025ac1f938a; detaching it from the instance and deleting it from the info cache [ 1167.508747] env[61243]: DEBUG nova.network.neutron [req-e1b5bd54-9ea3-45af-b531-e07d25477ee3 req-3a9c1aad-f757-4469-aa98-804407a82e54 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.899600] env[61243]: INFO nova.scheduler.client.report [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleted allocation for migration bc1e3cd7-8657-4772-903c-ca1056270095 [ 1167.988655] env[61243]: DEBUG nova.network.neutron [-] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1168.011395] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-801d4d7f-2add-48be-a971-ad3a06645df6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.021461] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b93dca5-b7c4-4109-9bee-53b9ee43c704 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.047910] env[61243]: DEBUG nova.compute.manager [req-e1b5bd54-9ea3-45af-b531-e07d25477ee3 req-3a9c1aad-f757-4469-aa98-804407a82e54 service nova] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Detach interface failed, port_id=00f9d661-7722-4165-8933-c025ac1f938a, reason: Instance 486ebf65-4376-424c-9deb-f1ef70e44be3 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1168.154274] env[61243]: INFO nova.compute.manager [None req-f6067bb5-c469-468a-aa52-5dfe388c2262 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Get console output [ 1168.154615] env[61243]: WARNING nova.virt.vmwareapi.driver [None req-f6067bb5-c469-468a-aa52-5dfe388c2262 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] The console log is missing. Check your VSPC configuration [ 1168.405721] env[61243]: DEBUG oslo_concurrency.lockutils [None req-a67ee9a2-b6cd-4b07-bdbe-7037a1df0307 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.639s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.490962] env[61243]: INFO nova.compute.manager [-] [instance: 486ebf65-4376-424c-9deb-f1ef70e44be3] Took 1.40 seconds to deallocate network for instance. [ 1168.997412] env[61243]: DEBUG oslo_concurrency.lockutils [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.997647] env[61243]: DEBUG oslo_concurrency.lockutils [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.997875] env[61243]: DEBUG nova.objects.instance [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lazy-loading 'resources' on Instance uuid 486ebf65-4376-424c-9deb-f1ef70e44be3 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1169.587924] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262782c3-5394-4a4b-ae2b-e70aa5baf17e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.596343] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf7a5f7-cbd5-4893-8859-1238f974598c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.627243] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07640a56-2a87-473d-a00f-d7c2ac03ff5d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.634554] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ebb0f4-85e5-48a9-9d4e-682c21c31f22 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.647397] env[61243]: DEBUG nova.compute.provider_tree [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1170.150790] env[61243]: DEBUG nova.scheduler.client.report [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1170.655806] env[61243]: DEBUG oslo_concurrency.lockutils [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.658s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.682583] env[61243]: INFO nova.scheduler.client.report [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Deleted allocations for instance 486ebf65-4376-424c-9deb-f1ef70e44be3 [ 1170.855445] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "440b35e6-eecb-4e20-831d-cf0db0158e81" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.855718] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.855905] env[61243]: INFO nova.compute.manager [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Shelving [ 1171.189429] env[61243]: DEBUG oslo_concurrency.lockutils [None req-335d5298-2eed-4ce5-9dd1-02f67feb7216 tempest-AttachVolumeShelveTestJSON-961961806 tempest-AttachVolumeShelveTestJSON-961961806-project-member] Lock "486ebf65-4376-424c-9deb-f1ef70e44be3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.835s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.364023] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1171.364465] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4e2d721e-ddd3-4581-9370-f71fd53495a3 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.371518] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1171.371518] env[61243]: value = "task-1339634" [ 1171.371518] env[61243]: _type = "Task" [ 1171.371518] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.379929] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339634, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.880798] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339634, 'name': PowerOffVM_Task, 'duration_secs': 0.177338} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.882022] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1171.882277] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e9c57f-c46a-4dcb-af57-b11dc44a5977 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.900663] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3f93ef-2156-4b0d-b2d4-83c9a467f523 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.410716] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Creating Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1172.411101] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8ddcfca0-d623-429e-a507-fd7f264fd0e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.419135] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1172.419135] env[61243]: value = "task-1339635" [ 1172.419135] env[61243]: _type = "Task" [ 1172.419135] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.427776] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339635, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.931584] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339635, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.432361] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339635, 'name': CreateSnapshot_Task, 'duration_secs': 0.693811} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.432629] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Created Snapshot of the VM instance {{(pid=61243) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1173.433394] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c0d12f-7958-4540-825b-98b2d6dcc412 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.951974] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Creating linked-clone VM from snapshot {{(pid=61243) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1173.951974] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9e2ba803-a4f2-42b0-bb09-518ef939745c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.959819] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1173.959819] env[61243]: value = "task-1339637" [ 1173.959819] env[61243]: _type = "Task" [ 1173.959819] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.967599] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339637, 'name': CloneVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.471015] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339637, 'name': CloneVM_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.971595] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339637, 'name': CloneVM_Task} progress is 95%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.470679] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339637, 'name': CloneVM_Task, 'duration_secs': 1.233015} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.470961] env[61243]: INFO nova.virt.vmwareapi.vmops [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Created linked-clone VM from snapshot [ 1175.471771] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d359e2-3e89-407e-8340-e35c5adccc1b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.478836] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Uploading image d3e17474-c388-4517-9d66-c712281235d1 {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1175.505357] env[61243]: DEBUG oslo_vmware.rw_handles [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1175.505357] env[61243]: value = "vm-285810" [ 1175.505357] env[61243]: _type = "VirtualMachine" [ 1175.505357] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1175.505685] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-47e4966d-84c5-46dc-a87f-1a1f877e4494 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.512596] env[61243]: DEBUG oslo_vmware.rw_handles [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lease: (returnval){ [ 1175.512596] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5267b426-caed-50ba-7af6-0ce41be7f8c6" [ 1175.512596] env[61243]: _type = "HttpNfcLease" [ 1175.512596] env[61243]: } obtained for exporting VM: (result){ [ 1175.512596] env[61243]: value = "vm-285810" [ 1175.512596] env[61243]: _type = "VirtualMachine" [ 1175.512596] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1175.512909] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the lease: (returnval){ [ 1175.512909] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5267b426-caed-50ba-7af6-0ce41be7f8c6" [ 1175.512909] env[61243]: _type = "HttpNfcLease" [ 1175.512909] env[61243]: } to be ready. {{(pid=61243) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1175.519096] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1175.519096] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5267b426-caed-50ba-7af6-0ce41be7f8c6" [ 1175.519096] env[61243]: _type = "HttpNfcLease" [ 1175.519096] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1176.020959] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1176.020959] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5267b426-caed-50ba-7af6-0ce41be7f8c6" [ 1176.020959] env[61243]: _type = "HttpNfcLease" [ 1176.020959] env[61243]: } is ready. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1176.021290] env[61243]: DEBUG oslo_vmware.rw_handles [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1176.021290] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]5267b426-caed-50ba-7af6-0ce41be7f8c6" [ 1176.021290] env[61243]: _type = "HttpNfcLease" [ 1176.021290] env[61243]: }. {{(pid=61243) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1176.021993] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29533e22-063c-4c4b-9dfd-90334bebc3d4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.029309] env[61243]: DEBUG oslo_vmware.rw_handles [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5271839a-0ef5-4322-39b6-c04b685051f3/disk-0.vmdk from lease info. {{(pid=61243) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1176.029494] env[61243]: DEBUG oslo_vmware.rw_handles [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5271839a-0ef5-4322-39b6-c04b685051f3/disk-0.vmdk for reading. {{(pid=61243) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1176.126717] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6ad09011-839e-4fe5-a558-bd288b95f2f6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.273541] env[61243]: DEBUG oslo_vmware.rw_handles [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5271839a-0ef5-4322-39b6-c04b685051f3/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1184.274471] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6c3c87f-b9cb-4a6b-9466-2135377ac461 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.282180] env[61243]: DEBUG oslo_vmware.rw_handles [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5271839a-0ef5-4322-39b6-c04b685051f3/disk-0.vmdk is in state: ready. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1184.282363] env[61243]: ERROR oslo_vmware.rw_handles [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5271839a-0ef5-4322-39b6-c04b685051f3/disk-0.vmdk due to incomplete transfer. [ 1184.282594] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-235ce0ac-b741-41be-a7e4-e0a9eceefa7d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.289695] env[61243]: DEBUG oslo_vmware.rw_handles [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5271839a-0ef5-4322-39b6-c04b685051f3/disk-0.vmdk. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1184.289904] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Uploaded image d3e17474-c388-4517-9d66-c712281235d1 to the Glance image server {{(pid=61243) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1184.292212] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Destroying the VM {{(pid=61243) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1184.292449] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e55ca956-9b0c-47d0-b3ee-98f9b468692f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.297530] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1184.297530] env[61243]: value = "task-1339639" [ 1184.297530] env[61243]: _type = "Task" [ 1184.297530] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.304825] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339639, 'name': Destroy_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.807305] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339639, 'name': Destroy_Task, 'duration_secs': 0.354537} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.807578] env[61243]: INFO nova.virt.vmwareapi.vm_util [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Destroyed the VM [ 1184.807819] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Deleting Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1184.808099] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5a07e923-d8cd-4725-9847-fc2f678b62cd {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.815022] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1184.815022] env[61243]: value = "task-1339640" [ 1184.815022] env[61243]: _type = "Task" [ 1184.815022] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.822668] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339640, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.325226] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339640, 'name': RemoveSnapshot_Task, 'duration_secs': 0.43256} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.325517] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Deleted Snapshot of the VM instance {{(pid=61243) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1185.325785] env[61243]: DEBUG nova.compute.manager [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1185.326545] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12798015-50ad-405e-b049-49fc53b661e1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.838379] env[61243]: INFO nova.compute.manager [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Shelve offloading [ 1185.840349] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1185.840625] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-745191e7-9133-4e71-ad8e-645db6e43c80 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.848184] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1185.848184] env[61243]: value = "task-1339641" [ 1185.848184] env[61243]: _type = "Task" [ 1185.848184] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.855786] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339641, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.358546] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] VM already powered off {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1186.358912] env[61243]: DEBUG nova.compute.manager [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1186.359505] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fba8e78-e419-4eca-a07a-12dd87d8016b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.364902] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1186.365090] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.365266] env[61243]: DEBUG nova.network.neutron [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1187.070611] env[61243]: DEBUG nova.network.neutron [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Updating instance_info_cache with network_info: [{"id": "723d362f-40a3-492a-9a70-f86c847d5d16", "address": "fa:16:3e:dc:f7:98", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap723d362f-40", "ovs_interfaceid": "723d362f-40a3-492a-9a70-f86c847d5d16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1187.573533] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1187.797827] env[61243]: DEBUG nova.compute.manager [req-20735e0c-1ac2-492c-9767-832191f8beaa req-44dcf21c-9bc8-4dbe-9c72-ddff80da4de3 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Received event network-vif-unplugged-723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1187.797990] env[61243]: DEBUG oslo_concurrency.lockutils [req-20735e0c-1ac2-492c-9767-832191f8beaa req-44dcf21c-9bc8-4dbe-9c72-ddff80da4de3 service nova] Acquiring lock "440b35e6-eecb-4e20-831d-cf0db0158e81-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.798367] env[61243]: DEBUG oslo_concurrency.lockutils [req-20735e0c-1ac2-492c-9767-832191f8beaa req-44dcf21c-9bc8-4dbe-9c72-ddff80da4de3 service nova] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.798563] env[61243]: DEBUG oslo_concurrency.lockutils [req-20735e0c-1ac2-492c-9767-832191f8beaa req-44dcf21c-9bc8-4dbe-9c72-ddff80da4de3 service nova] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.798742] env[61243]: DEBUG nova.compute.manager [req-20735e0c-1ac2-492c-9767-832191f8beaa req-44dcf21c-9bc8-4dbe-9c72-ddff80da4de3 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] No waiting events found dispatching network-vif-unplugged-723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1187.799022] env[61243]: WARNING nova.compute.manager [req-20735e0c-1ac2-492c-9767-832191f8beaa req-44dcf21c-9bc8-4dbe-9c72-ddff80da4de3 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Received unexpected event network-vif-unplugged-723d362f-40a3-492a-9a70-f86c847d5d16 for instance with vm_state shelved and task_state shelving_offloading. [ 1187.893177] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1187.894136] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf48c13b-47f3-49c0-8044-c17aee3c3eaa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.901642] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1187.901895] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-14140544-94e9-4a8a-a8e2-2e8c5fab219e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.966066] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1187.966378] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1187.966639] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Deleting the datastore file [datastore2] 440b35e6-eecb-4e20-831d-cf0db0158e81 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1187.966968] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-899af104-c98e-4368-a6a0-818b999d7467 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.973521] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1187.973521] env[61243]: value = "task-1339643" [ 1187.973521] env[61243]: _type = "Task" [ 1187.973521] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.981467] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339643, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.483870] env[61243]: DEBUG oslo_vmware.api [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339643, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135455} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.484127] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1188.484325] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1188.484506] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1188.512665] env[61243]: INFO nova.scheduler.client.report [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Deleted allocations for instance 440b35e6-eecb-4e20-831d-cf0db0158e81 [ 1189.017326] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.017685] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.017939] env[61243]: DEBUG nova.objects.instance [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lazy-loading 'resources' on Instance uuid 440b35e6-eecb-4e20-831d-cf0db0158e81 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1189.520573] env[61243]: DEBUG nova.objects.instance [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lazy-loading 'numa_topology' on Instance uuid 440b35e6-eecb-4e20-831d-cf0db0158e81 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1189.823794] env[61243]: DEBUG nova.compute.manager [req-cbdbe531-b0a0-4923-918d-d5c13c098c20 req-34ee760b-2319-4cfd-8b6b-f41b415d435d service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Received event network-changed-723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1189.824017] env[61243]: DEBUG nova.compute.manager [req-cbdbe531-b0a0-4923-918d-d5c13c098c20 req-34ee760b-2319-4cfd-8b6b-f41b415d435d service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Refreshing instance network info cache due to event network-changed-723d362f-40a3-492a-9a70-f86c847d5d16. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1189.824242] env[61243]: DEBUG oslo_concurrency.lockutils [req-cbdbe531-b0a0-4923-918d-d5c13c098c20 req-34ee760b-2319-4cfd-8b6b-f41b415d435d service nova] Acquiring lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.824395] env[61243]: DEBUG oslo_concurrency.lockutils [req-cbdbe531-b0a0-4923-918d-d5c13c098c20 req-34ee760b-2319-4cfd-8b6b-f41b415d435d service nova] Acquired lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.824571] env[61243]: DEBUG nova.network.neutron [req-cbdbe531-b0a0-4923-918d-d5c13c098c20 req-34ee760b-2319-4cfd-8b6b-f41b415d435d service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Refreshing network info cache for port 723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1190.023419] env[61243]: DEBUG nova.objects.base [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Object Instance<440b35e6-eecb-4e20-831d-cf0db0158e81> lazy-loaded attributes: resources,numa_topology {{(pid=61243) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1190.075488] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d3c7c13-e050-4b38-bb6a-fc5c815a01e2 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.082855] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a33c00-5786-4a59-b183-7ac197f69aaf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.112758] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9daf0ddb-f8b3-442b-808a-5641657f9233 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.120458] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c63ebdcd-fa84-4cde-bf2a-317d4ecabacb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.133260] env[61243]: DEBUG nova.compute.provider_tree [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1190.545070] env[61243]: DEBUG nova.network.neutron [req-cbdbe531-b0a0-4923-918d-d5c13c098c20 req-34ee760b-2319-4cfd-8b6b-f41b415d435d service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Updated VIF entry in instance network info cache for port 723d362f-40a3-492a-9a70-f86c847d5d16. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1190.545445] env[61243]: DEBUG nova.network.neutron [req-cbdbe531-b0a0-4923-918d-d5c13c098c20 req-34ee760b-2319-4cfd-8b6b-f41b415d435d service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Updating instance_info_cache with network_info: [{"id": "723d362f-40a3-492a-9a70-f86c847d5d16", "address": "fa:16:3e:dc:f7:98", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": null, "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap723d362f-40", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1190.637487] env[61243]: DEBUG nova.scheduler.client.report [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1190.835591] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "440b35e6-eecb-4e20-831d-cf0db0158e81" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.048867] env[61243]: DEBUG oslo_concurrency.lockutils [req-cbdbe531-b0a0-4923-918d-d5c13c098c20 req-34ee760b-2319-4cfd-8b6b-f41b415d435d service nova] Releasing lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.142316] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.124s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.649539] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d0f0f20b-c825-4d28-a65e-583b35639c1f tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.794s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.650389] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.815s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.650580] env[61243]: INFO nova.compute.manager [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Unshelving [ 1192.673548] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1192.673842] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1192.674077] env[61243]: DEBUG nova.objects.instance [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lazy-loading 'pci_requests' on Instance uuid 440b35e6-eecb-4e20-831d-cf0db0158e81 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1193.178243] env[61243]: DEBUG nova.objects.instance [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lazy-loading 'numa_topology' on Instance uuid 440b35e6-eecb-4e20-831d-cf0db0158e81 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1193.302064] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.681293] env[61243]: INFO nova.compute.claims [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1194.747216] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34436a41-c712-4dc9-9f0a-46a472f7f704 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.754358] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57fe237a-aa16-43c5-9fb1-3f1d8d48cc6f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.782741] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d35127-82e9-480f-ae59-8f77e9651d4d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.789359] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1184883-0c7b-4578-939b-b72ba4e8060a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.802389] env[61243]: DEBUG nova.compute.provider_tree [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1195.097664] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.097920] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.098158] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.098353] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.098529] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.100955] env[61243]: INFO nova.compute.manager [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Terminating instance [ 1195.102811] env[61243]: DEBUG nova.compute.manager [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1195.103049] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1195.103293] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0834366b-11b2-41cd-a0a0-4e6f49f339a9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.110149] env[61243]: DEBUG oslo_vmware.api [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1195.110149] env[61243]: value = "task-1339644" [ 1195.110149] env[61243]: _type = "Task" [ 1195.110149] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.117822] env[61243]: DEBUG oslo_vmware.api [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339644, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.301777] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1195.302055] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1195.305081] env[61243]: DEBUG nova.scheduler.client.report [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1195.619943] env[61243]: DEBUG oslo_vmware.api [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339644, 'name': PowerOffVM_Task, 'duration_secs': 0.206147} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.620185] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1195.620391] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Volume detach. Driver type: vmdk {{(pid=61243) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1195.620596] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285805', 'volume_id': 'd7581ea4-bb02-421e-a48c-c34488368a12', 'name': 'volume-d7581ea4-bb02-421e-a48c-c34488368a12', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4', 'attached_at': '2024-10-22T18:14:26.000000', 'detached_at': '', 'volume_id': 'd7581ea4-bb02-421e-a48c-c34488368a12', 'serial': 'd7581ea4-bb02-421e-a48c-c34488368a12'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1195.621327] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f05aff-2362-4ce3-8acb-787db030d6db {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.639831] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e1d90e6-3c9f-4214-9699-56f33e5bc073 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.645911] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030e0117-2ed0-4df9-a6be-352bd2b2e1a9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.662201] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c79a882-424f-4ba1-9771-572bf9144d7a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.675929] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] The volume has not been displaced from its original location: [datastore2] volume-d7581ea4-bb02-421e-a48c-c34488368a12/volume-d7581ea4-bb02-421e-a48c-c34488368a12.vmdk. No consolidation needed. {{(pid=61243) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1195.680942] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Reconfiguring VM instance instance-00000071 to detach disk 2000 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1195.681210] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8ce2c2a-b68f-4eb4-bad9-94958f1accdf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.697236] env[61243]: DEBUG oslo_vmware.api [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1195.697236] env[61243]: value = "task-1339645" [ 1195.697236] env[61243]: _type = "Task" [ 1195.697236] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.704340] env[61243]: DEBUG oslo_vmware.api [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339645, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.809611] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.136s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.008487] env[61243]: INFO nova.network.neutron [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Updating port 723d362f-40a3-492a-9a70-f86c847d5d16 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1196.206861] env[61243]: DEBUG oslo_vmware.api [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339645, 'name': ReconfigVM_Task, 'duration_secs': 0.169012} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.208253] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Reconfigured VM instance instance-00000071 to detach disk 2000 {{(pid=61243) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1196.211957] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c55615c0-c9a8-48a6-8dc9-5ca3a2e6b861 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.226867] env[61243]: DEBUG oslo_vmware.api [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1196.226867] env[61243]: value = "task-1339646" [ 1196.226867] env[61243]: _type = "Task" [ 1196.226867] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.237463] env[61243]: DEBUG oslo_vmware.api [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339646, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.302210] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.302446] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager.update_available_resource {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.736414] env[61243]: DEBUG oslo_vmware.api [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339646, 'name': ReconfigVM_Task, 'duration_secs': 0.170616} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.736769] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285805', 'volume_id': 'd7581ea4-bb02-421e-a48c-c34488368a12', 'name': 'volume-d7581ea4-bb02-421e-a48c-c34488368a12', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4', 'attached_at': '2024-10-22T18:14:26.000000', 'detached_at': '', 'volume_id': 'd7581ea4-bb02-421e-a48c-c34488368a12', 'serial': 'd7581ea4-bb02-421e-a48c-c34488368a12'} {{(pid=61243) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1196.737083] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1196.737847] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64200d7e-ce40-4a79-ac26-b2d6f062d420 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.744109] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1196.744331] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd299c7f-ab39-466f-869d-62a1fe4fd614 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.805141] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.805396] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.805578] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.805741] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61243) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1196.806654] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-978e1625-8651-4a2a-b897-cb0a222b7f41 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.813136] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1196.813476] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1196.813476] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleting the datastore file [datastore2] 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1196.813722] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9029d392-5e9c-493a-8f4a-c41c441b5ad1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.820389] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2aa200-0ba0-4ec1-a6f4-0fc095f3ff21 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.825948] env[61243]: DEBUG oslo_vmware.api [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1196.825948] env[61243]: value = "task-1339648" [ 1196.825948] env[61243]: _type = "Task" [ 1196.825948] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.839526] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fadf5ee-fb38-4cf0-ad06-67985e5d3c47 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.844822] env[61243]: DEBUG oslo_vmware.api [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339648, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.848960] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d01978b9-c0ba-4bff-a856-02093376693d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.877253] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180498MB free_disk=154GB free_vcpus=48 pci_devices=None {{(pid=61243) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1196.877406] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.877605] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.335971] env[61243]: DEBUG oslo_vmware.api [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339648, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144725} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.336256] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1197.336454] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1197.336640] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1197.336820] env[61243]: INFO nova.compute.manager [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Took 2.23 seconds to destroy the instance on the hypervisor. [ 1197.337073] env[61243]: DEBUG oslo.service.loopingcall [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1197.337271] env[61243]: DEBUG nova.compute.manager [-] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1197.337369] env[61243]: DEBUG nova.network.neutron [-] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1197.460967] env[61243]: DEBUG nova.compute.manager [req-b2a6e673-9a47-4d1b-989c-511633657d86 req-7daba22f-5670-4de8-90a8-13bda1fc5415 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Received event network-vif-plugged-723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1197.461206] env[61243]: DEBUG oslo_concurrency.lockutils [req-b2a6e673-9a47-4d1b-989c-511633657d86 req-7daba22f-5670-4de8-90a8-13bda1fc5415 service nova] Acquiring lock "440b35e6-eecb-4e20-831d-cf0db0158e81-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.461423] env[61243]: DEBUG oslo_concurrency.lockutils [req-b2a6e673-9a47-4d1b-989c-511633657d86 req-7daba22f-5670-4de8-90a8-13bda1fc5415 service nova] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.461646] env[61243]: DEBUG oslo_concurrency.lockutils [req-b2a6e673-9a47-4d1b-989c-511633657d86 req-7daba22f-5670-4de8-90a8-13bda1fc5415 service nova] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.461841] env[61243]: DEBUG nova.compute.manager [req-b2a6e673-9a47-4d1b-989c-511633657d86 req-7daba22f-5670-4de8-90a8-13bda1fc5415 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] No waiting events found dispatching network-vif-plugged-723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1197.462021] env[61243]: WARNING nova.compute.manager [req-b2a6e673-9a47-4d1b-989c-511633657d86 req-7daba22f-5670-4de8-90a8-13bda1fc5415 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Received unexpected event network-vif-plugged-723d362f-40a3-492a-9a70-f86c847d5d16 for instance with vm_state shelved_offloaded and task_state spawning. [ 1197.638594] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1197.638594] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1197.638745] env[61243]: DEBUG nova.network.neutron [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Building network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1197.905514] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 3f839aa5-2b9a-4807-b63b-931f74455532 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1197.905797] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 91b14a22-811a-4151-9769-9c684ec993db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1197.905797] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1197.905904] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Instance 440b35e6-eecb-4e20-831d-cf0db0158e81 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61243) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1197.906103] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1197.906249] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1344MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61243) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1197.962310] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc36d80-bd84-48b6-8fae-1251aad975f9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.970103] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb771622-f469-4138-9851-f1dbd38fd040 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.001037] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd43b30-5846-4066-a41c-02477600113f {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.008077] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8744b4a-272d-4183-b557-8ae186f5e7b6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.021022] env[61243]: DEBUG nova.compute.provider_tree [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1198.350522] env[61243]: DEBUG nova.network.neutron [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Updating instance_info_cache with network_info: [{"id": "723d362f-40a3-492a-9a70-f86c847d5d16", "address": "fa:16:3e:dc:f7:98", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap723d362f-40", "ovs_interfaceid": "723d362f-40a3-492a-9a70-f86c847d5d16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.391141] env[61243]: DEBUG nova.network.neutron [-] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.524502] env[61243]: DEBUG nova.scheduler.client.report [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1198.853514] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1198.880066] env[61243]: DEBUG nova.virt.hardware [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T18:02:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='c0232591ff8cdc0882eb02c63b333d85',container_format='bare',created_at=2024-10-22T18:14:39Z,direct_url=,disk_format='vmdk',id=d3e17474-c388-4517-9d66-c712281235d1,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1944516553-shelved',owner='66e2ba61b3354ef1b496232264d27600',properties=ImageMetaProps,protected=,size=31663104,status='active',tags=,updated_at=2024-10-22T18:14:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1198.880336] env[61243]: DEBUG nova.virt.hardware [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1198.880505] env[61243]: DEBUG nova.virt.hardware [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image limits 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1198.880693] env[61243]: DEBUG nova.virt.hardware [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Flavor pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1198.880851] env[61243]: DEBUG nova.virt.hardware [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Image pref 0:0:0 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1198.881014] env[61243]: DEBUG nova.virt.hardware [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61243) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1198.881237] env[61243]: DEBUG nova.virt.hardware [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1198.881651] env[61243]: DEBUG nova.virt.hardware [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1198.881861] env[61243]: DEBUG nova.virt.hardware [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Got 1 possible topologies {{(pid=61243) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1198.882054] env[61243]: DEBUG nova.virt.hardware [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1198.882241] env[61243]: DEBUG nova.virt.hardware [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61243) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1198.883112] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1518a82e-ae15-416d-b632-e4b5b84b2451 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.891299] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0057fb6-107e-4944-92ca-2989a0cdef3c {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.894913] env[61243]: INFO nova.compute.manager [-] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Took 1.56 seconds to deallocate network for instance. [ 1198.906179] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:f7:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd7b5f1ef-d4b9-4ec3-b047-17e4cb349d25', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '723d362f-40a3-492a-9a70-f86c847d5d16', 'vif_model': 'vmxnet3'}] {{(pid=61243) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1198.913559] env[61243]: DEBUG oslo.service.loopingcall [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1198.913646] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Creating VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1198.913872] env[61243]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-952aeac7-e13e-4ba4-ac3a-b00db05352a9 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.934673] env[61243]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1198.934673] env[61243]: value = "task-1339649" [ 1198.934673] env[61243]: _type = "Task" [ 1198.934673] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.943014] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339649, 'name': CreateVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.029158] env[61243]: DEBUG nova.compute.resource_tracker [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61243) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1199.029391] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.152s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1199.440573] env[61243]: INFO nova.compute.manager [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Took 0.55 seconds to detach 1 volumes for instance. [ 1199.445455] env[61243]: DEBUG oslo_vmware.api [-] Task: {'id': task-1339649, 'name': CreateVM_Task, 'duration_secs': 0.29558} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.445938] env[61243]: DEBUG nova.compute.manager [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Deleting volume: d7581ea4-bb02-421e-a48c-c34488368a12 {{(pid=61243) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1199.447659] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Created VM on the ESX host {{(pid=61243) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1199.448415] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d3e17474-c388-4517-9d66-c712281235d1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.448605] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d3e17474-c388-4517-9d66-c712281235d1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.448972] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d3e17474-c388-4517-9d66-c712281235d1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1199.449232] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d4ad3ec-4157-46d7-929e-2d6c5cb933e8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.453616] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1199.453616] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f57503-8fac-cbd9-6044-9098f5118e30" [ 1199.453616] env[61243]: _type = "Task" [ 1199.453616] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.460872] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52f57503-8fac-cbd9-6044-9098f5118e30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.490709] env[61243]: DEBUG nova.compute.manager [req-006b1d9f-c684-4bce-af9c-196abefaee7f req-332f9c23-103b-4144-b248-6d73c057f18c service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Received event network-changed-723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1199.490948] env[61243]: DEBUG nova.compute.manager [req-006b1d9f-c684-4bce-af9c-196abefaee7f req-332f9c23-103b-4144-b248-6d73c057f18c service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Refreshing instance network info cache due to event network-changed-723d362f-40a3-492a-9a70-f86c847d5d16. {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1199.491340] env[61243]: DEBUG oslo_concurrency.lockutils [req-006b1d9f-c684-4bce-af9c-196abefaee7f req-332f9c23-103b-4144-b248-6d73c057f18c service nova] Acquiring lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.491340] env[61243]: DEBUG oslo_concurrency.lockutils [req-006b1d9f-c684-4bce-af9c-196abefaee7f req-332f9c23-103b-4144-b248-6d73c057f18c service nova] Acquired lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.491453] env[61243]: DEBUG nova.network.neutron [req-006b1d9f-c684-4bce-af9c-196abefaee7f req-332f9c23-103b-4144-b248-6d73c057f18c service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Refreshing network info cache for port 723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1199.967019] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d3e17474-c388-4517-9d66-c712281235d1" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1199.967019] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Processing image d3e17474-c388-4517-9d66-c712281235d1 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1199.967019] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d3e17474-c388-4517-9d66-c712281235d1/d3e17474-c388-4517-9d66-c712281235d1.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.967019] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d3e17474-c388-4517-9d66-c712281235d1/d3e17474-c388-4517-9d66-c712281235d1.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.967019] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1199.967019] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9b465152-a330-418b-9e10-49ee1a3f0440 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.973531] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1199.973865] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61243) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1199.974628] env[61243]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-897d078c-8599-4438-9899-88e82d7a3b1b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.979119] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1199.979119] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52171a70-e0dc-2057-9878-9d26b1c5105d" [ 1199.979119] env[61243]: _type = "Task" [ 1199.979119] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.986698] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1199.987053] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1199.987393] env[61243]: DEBUG nova.objects.instance [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lazy-loading 'resources' on Instance uuid 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1199.988489] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]52171a70-e0dc-2057-9878-9d26b1c5105d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.190516] env[61243]: DEBUG nova.network.neutron [req-006b1d9f-c684-4bce-af9c-196abefaee7f req-332f9c23-103b-4144-b248-6d73c057f18c service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Updated VIF entry in instance network info cache for port 723d362f-40a3-492a-9a70-f86c847d5d16. {{(pid=61243) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1200.190932] env[61243]: DEBUG nova.network.neutron [req-006b1d9f-c684-4bce-af9c-196abefaee7f req-332f9c23-103b-4144-b248-6d73c057f18c service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Updating instance_info_cache with network_info: [{"id": "723d362f-40a3-492a-9a70-f86c847d5d16", "address": "fa:16:3e:dc:f7:98", "network": {"id": "83cb5daf-0ded-4b94-8d5b-da177f7fbbe1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-477087032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66e2ba61b3354ef1b496232264d27600", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap723d362f-40", "ovs_interfaceid": "723d362f-40a3-492a-9a70-f86c847d5d16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1200.492543] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Preparing fetch location {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1200.492837] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Fetch image to [datastore2] OSTACK_IMG_127f28f7-1e9d-4651-9b1d-14af247c5645/OSTACK_IMG_127f28f7-1e9d-4651-9b1d-14af247c5645.vmdk {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1200.493118] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Downloading stream optimized image d3e17474-c388-4517-9d66-c712281235d1 to [datastore2] OSTACK_IMG_127f28f7-1e9d-4651-9b1d-14af247c5645/OSTACK_IMG_127f28f7-1e9d-4651-9b1d-14af247c5645.vmdk on the data store datastore2 as vApp {{(pid=61243) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1200.493213] env[61243]: DEBUG nova.virt.vmwareapi.images [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Downloading image file data d3e17474-c388-4517-9d66-c712281235d1 to the ESX as VM named 'OSTACK_IMG_127f28f7-1e9d-4651-9b1d-14af247c5645' {{(pid=61243) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1200.562777] env[61243]: DEBUG oslo_vmware.rw_handles [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1200.562777] env[61243]: value = "resgroup-9" [ 1200.562777] env[61243]: _type = "ResourcePool" [ 1200.562777] env[61243]: }. {{(pid=61243) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1200.563086] env[61243]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-6cab1327-0b37-43ab-a15d-b7f601ae7b85 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.578779] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0c28bf-93f7-4421-9933-6f3474e6fd70 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.585724] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb8420f-fdbd-4f39-a0f8-5d4e628e25cb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.589612] env[61243]: DEBUG oslo_vmware.rw_handles [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lease: (returnval){ [ 1200.589612] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521607cb-2a59-1b9b-dc36-4a2b612677fc" [ 1200.589612] env[61243]: _type = "HttpNfcLease" [ 1200.589612] env[61243]: } obtained for vApp import into resource pool (val){ [ 1200.589612] env[61243]: value = "resgroup-9" [ 1200.589612] env[61243]: _type = "ResourcePool" [ 1200.589612] env[61243]: }. {{(pid=61243) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1200.589885] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the lease: (returnval){ [ 1200.589885] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521607cb-2a59-1b9b-dc36-4a2b612677fc" [ 1200.589885] env[61243]: _type = "HttpNfcLease" [ 1200.589885] env[61243]: } to be ready. {{(pid=61243) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1200.618045] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f710ce-182c-4749-a222-b1d73720f4d4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.622120] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1200.622120] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521607cb-2a59-1b9b-dc36-4a2b612677fc" [ 1200.622120] env[61243]: _type = "HttpNfcLease" [ 1200.622120] env[61243]: } is initializing. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1200.626717] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e4c533-e026-4a82-a1c4-c3f547beae46 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.640548] env[61243]: DEBUG nova.compute.provider_tree [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1200.693308] env[61243]: DEBUG oslo_concurrency.lockutils [req-006b1d9f-c684-4bce-af9c-196abefaee7f req-332f9c23-103b-4144-b248-6d73c057f18c service nova] Releasing lock "refresh_cache-440b35e6-eecb-4e20-831d-cf0db0158e81" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1200.693570] env[61243]: DEBUG nova.compute.manager [req-006b1d9f-c684-4bce-af9c-196abefaee7f req-332f9c23-103b-4144-b248-6d73c057f18c service nova] [instance: 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4] Received event network-vif-deleted-2168e525-2705-4de2-9879-ac1a43eb9f84 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1201.024299] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.024602] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.098349] env[61243]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1201.098349] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521607cb-2a59-1b9b-dc36-4a2b612677fc" [ 1201.098349] env[61243]: _type = "HttpNfcLease" [ 1201.098349] env[61243]: } is ready. {{(pid=61243) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1201.098653] env[61243]: DEBUG oslo_vmware.rw_handles [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1201.098653] env[61243]: value = "session[52af8c46-959e-dd7f-58ff-3f3e11362f5b]521607cb-2a59-1b9b-dc36-4a2b612677fc" [ 1201.098653] env[61243]: _type = "HttpNfcLease" [ 1201.098653] env[61243]: }. {{(pid=61243) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1201.099383] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1841b1d-be7d-4e94-a20a-a277392318f4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.107062] env[61243]: DEBUG oslo_vmware.rw_handles [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52399f08-e45d-89b9-6510-c9e62b4c1d5f/disk-0.vmdk from lease info. {{(pid=61243) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1201.107302] env[61243]: DEBUG oslo_vmware.rw_handles [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Creating HTTP connection to write to file with size = 31663104 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52399f08-e45d-89b9-6510-c9e62b4c1d5f/disk-0.vmdk. {{(pid=61243) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1201.163729] env[61243]: DEBUG nova.scheduler.client.report [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1201.171461] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c928aa84-a7cd-4c3c-88de-b3eba20ec1cf {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.529621] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.529941] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Starting heal instance info cache {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1201.669673] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.682s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1201.690875] env[61243]: INFO nova.scheduler.client.report [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleted allocations for instance 9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4 [ 1202.200330] env[61243]: DEBUG oslo_concurrency.lockutils [None req-2d370bf0-d08b-418f-90e3-592d54db80ed tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "9fc84e72-2ac8-476f-ad00-8e3bd8dc80e4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.102s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.379794] env[61243]: DEBUG oslo_vmware.rw_handles [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Completed reading data from the image iterator. {{(pid=61243) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1202.380033] env[61243]: DEBUG oslo_vmware.rw_handles [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52399f08-e45d-89b9-6510-c9e62b4c1d5f/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1202.380946] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55138631-d5ee-4852-8f4b-984a8879e368 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.387750] env[61243]: DEBUG oslo_vmware.rw_handles [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52399f08-e45d-89b9-6510-c9e62b4c1d5f/disk-0.vmdk is in state: ready. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1202.387941] env[61243]: DEBUG oslo_vmware.rw_handles [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52399f08-e45d-89b9-6510-c9e62b4c1d5f/disk-0.vmdk. {{(pid=61243) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1202.388226] env[61243]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-25f20eb8-e667-457b-a750-f522a64c1ae8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.563933] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquiring lock "refresh_cache-91b14a22-811a-4151-9769-9c684ec993db" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1202.564129] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Acquired lock "refresh_cache-91b14a22-811a-4151-9769-9c684ec993db" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1202.564250] env[61243]: DEBUG nova.network.neutron [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Forcefully refreshing network info cache for instance {{(pid=61243) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1202.570923] env[61243]: DEBUG oslo_vmware.rw_handles [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52399f08-e45d-89b9-6510-c9e62b4c1d5f/disk-0.vmdk. {{(pid=61243) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1202.571163] env[61243]: INFO nova.virt.vmwareapi.images [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Downloaded image file data d3e17474-c388-4517-9d66-c712281235d1 [ 1202.572012] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fbda046-eef6-41ee-81f2-ae8e848d56ca {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.587412] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-55032633-dd5a-4aec-94f2-6db711cc7f13 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.648513] env[61243]: INFO nova.virt.vmwareapi.images [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] The imported VM was unregistered [ 1202.650744] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Caching image {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1202.650980] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Creating directory with path [datastore2] devstack-image-cache_base/d3e17474-c388-4517-9d66-c712281235d1 {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1202.651505] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4cf5afac-97f2-4155-a9e6-45f77e5cf3c1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.661318] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Created directory with path [datastore2] devstack-image-cache_base/d3e17474-c388-4517-9d66-c712281235d1 {{(pid=61243) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1202.661502] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_127f28f7-1e9d-4651-9b1d-14af247c5645/OSTACK_IMG_127f28f7-1e9d-4651-9b1d-14af247c5645.vmdk to [datastore2] devstack-image-cache_base/d3e17474-c388-4517-9d66-c712281235d1/d3e17474-c388-4517-9d66-c712281235d1.vmdk. {{(pid=61243) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1202.661757] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-5425ab40-9bbd-4a20-8dac-150090467df8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.667838] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1202.667838] env[61243]: value = "task-1339653" [ 1202.667838] env[61243]: _type = "Task" [ 1202.667838] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.674673] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339653, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.771152] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "91b14a22-811a-4151-9769-9c684ec993db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1202.771474] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "91b14a22-811a-4151-9769-9c684ec993db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1202.771735] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "91b14a22-811a-4151-9769-9c684ec993db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1202.771967] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "91b14a22-811a-4151-9769-9c684ec993db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1202.772151] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "91b14a22-811a-4151-9769-9c684ec993db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.774377] env[61243]: INFO nova.compute.manager [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Terminating instance [ 1202.776189] env[61243]: DEBUG nova.compute.manager [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1202.776417] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1202.777264] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373aa8d8-bcc1-4f38-aa05-ce68dd14f89a {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.784697] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1202.784941] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-81ffd71a-23ea-4071-a478-a3eee8c7ed8e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.789822] env[61243]: DEBUG oslo_vmware.api [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1202.789822] env[61243]: value = "task-1339654" [ 1202.789822] env[61243]: _type = "Task" [ 1202.789822] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.797165] env[61243]: DEBUG oslo_vmware.api [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339654, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.178827] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339653, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.300771] env[61243]: DEBUG oslo_vmware.api [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339654, 'name': PowerOffVM_Task, 'duration_secs': 0.262341} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.301161] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1203.301253] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1203.301591] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-090c650d-6c0f-4bc8-845a-93062b4b3e9b {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.374955] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1203.375241] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1203.375451] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleting the datastore file [datastore2] 91b14a22-811a-4151-9769-9c684ec993db {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1203.375727] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-86039258-6a94-4556-92c7-537343dbbdab {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.382860] env[61243]: DEBUG oslo_vmware.api [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1203.382860] env[61243]: value = "task-1339656" [ 1203.382860] env[61243]: _type = "Task" [ 1203.382860] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.391850] env[61243]: DEBUG oslo_vmware.api [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339656, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.678631] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339653, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.787520] env[61243]: DEBUG nova.network.neutron [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Updating instance_info_cache with network_info: [{"id": "41768d32-20ce-414e-9c85-be16e03d795a", "address": "fa:16:3e:3c:5e:26", "network": {"id": "372599d0-3518-484c-952f-8ee192a8a8bf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1263099881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "324ed556b6f5419fa0376a7cfc508827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff3ecd2f-0b10-4faf-a512-fd7a20c28df1", "external-id": "nsx-vlan-transportzone-291", "segmentation_id": 291, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41768d32-20", "ovs_interfaceid": "41768d32-20ce-414e-9c85-be16e03d795a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1203.894676] env[61243]: DEBUG oslo_vmware.api [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339656, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.181655] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339653, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.290731] env[61243]: DEBUG oslo_concurrency.lockutils [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Releasing lock "refresh_cache-91b14a22-811a-4151-9769-9c684ec993db" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1204.290995] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Updated the network info_cache for instance {{(pid=61243) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1204.291207] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1204.291438] env[61243]: DEBUG oslo_service.periodic_task [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61243) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1204.291611] env[61243]: DEBUG nova.compute.manager [None req-d2208d62-ad26-4868-9f86-4ff8d420dead None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61243) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1204.393757] env[61243]: DEBUG oslo_vmware.api [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339656, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.680542] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339653, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.893815] env[61243]: DEBUG oslo_vmware.api [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339656, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.181512] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339653, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.394561] env[61243]: DEBUG oslo_vmware.api [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339656, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.666092} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.394897] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1205.395032] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1205.395229] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1205.395411] env[61243]: INFO nova.compute.manager [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Took 2.62 seconds to destroy the instance on the hypervisor. [ 1205.395661] env[61243]: DEBUG oslo.service.loopingcall [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1205.395863] env[61243]: DEBUG nova.compute.manager [-] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1205.395960] env[61243]: DEBUG nova.network.neutron [-] [instance: 91b14a22-811a-4151-9769-9c684ec993db] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1205.644542] env[61243]: DEBUG nova.compute.manager [req-3564b67d-4bb6-46eb-b9ff-e82221059dcc req-acca6024-d0ca-4c7e-844a-7b990f03c112 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Received event network-vif-deleted-41768d32-20ce-414e-9c85-be16e03d795a {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1205.644730] env[61243]: INFO nova.compute.manager [req-3564b67d-4bb6-46eb-b9ff-e82221059dcc req-acca6024-d0ca-4c7e-844a-7b990f03c112 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Neutron deleted interface 41768d32-20ce-414e-9c85-be16e03d795a; detaching it from the instance and deleting it from the info cache [ 1205.644918] env[61243]: DEBUG nova.network.neutron [req-3564b67d-4bb6-46eb-b9ff-e82221059dcc req-acca6024-d0ca-4c7e-844a-7b990f03c112 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1205.681955] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339653, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.581096} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.682224] env[61243]: INFO nova.virt.vmwareapi.ds_util [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_127f28f7-1e9d-4651-9b1d-14af247c5645/OSTACK_IMG_127f28f7-1e9d-4651-9b1d-14af247c5645.vmdk to [datastore2] devstack-image-cache_base/d3e17474-c388-4517-9d66-c712281235d1/d3e17474-c388-4517-9d66-c712281235d1.vmdk. [ 1205.682413] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Cleaning up location [datastore2] OSTACK_IMG_127f28f7-1e9d-4651-9b1d-14af247c5645 {{(pid=61243) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1205.682595] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_127f28f7-1e9d-4651-9b1d-14af247c5645 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1205.682847] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-854eb353-1244-497f-a794-26f260424a31 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.688948] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1205.688948] env[61243]: value = "task-1339657" [ 1205.688948] env[61243]: _type = "Task" [ 1205.688948] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.696457] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339657, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.125822] env[61243]: DEBUG nova.network.neutron [-] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1206.147347] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c0b1edf8-2421-4435-82de-9403e19f67c6 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.156713] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c82b9be-680b-4964-81ff-499c540ed1ae {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.181354] env[61243]: DEBUG nova.compute.manager [req-3564b67d-4bb6-46eb-b9ff-e82221059dcc req-acca6024-d0ca-4c7e-844a-7b990f03c112 service nova] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Detach interface failed, port_id=41768d32-20ce-414e-9c85-be16e03d795a, reason: Instance 91b14a22-811a-4151-9769-9c684ec993db could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1206.196913] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339657, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.035615} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.197164] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1206.197337] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d3e17474-c388-4517-9d66-c712281235d1/d3e17474-c388-4517-9d66-c712281235d1.vmdk" {{(pid=61243) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1206.197578] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d3e17474-c388-4517-9d66-c712281235d1/d3e17474-c388-4517-9d66-c712281235d1.vmdk to [datastore2] 440b35e6-eecb-4e20-831d-cf0db0158e81/440b35e6-eecb-4e20-831d-cf0db0158e81.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1206.197810] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32999a7d-ee16-48f8-89fb-5e3b47dd6e67 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.203707] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1206.203707] env[61243]: value = "task-1339658" [ 1206.203707] env[61243]: _type = "Task" [ 1206.203707] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.210396] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339658, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.629187] env[61243]: INFO nova.compute.manager [-] [instance: 91b14a22-811a-4151-9769-9c684ec993db] Took 1.23 seconds to deallocate network for instance. [ 1206.714179] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339658, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.136667] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.136936] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.137181] env[61243]: DEBUG nova.objects.instance [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lazy-loading 'resources' on Instance uuid 91b14a22-811a-4151-9769-9c684ec993db {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1207.215150] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339658, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.700563] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2dd6e49-52b2-4f8a-9389-ba46ab258ba8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.711916] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a50d2833-2a29-4050-a075-0d9cf2ae86b8 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.720102] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339658, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.744012] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d45981-5a96-4b6c-8abc-916a0466b58e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.751546] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a24c27e-ccf8-4127-a824-61c0a74ce39d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.766366] env[61243]: DEBUG nova.compute.provider_tree [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1208.218186] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339658, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.269543] env[61243]: DEBUG nova.scheduler.client.report [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1208.719251] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339658, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.133821} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.719723] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d3e17474-c388-4517-9d66-c712281235d1/d3e17474-c388-4517-9d66-c712281235d1.vmdk to [datastore2] 440b35e6-eecb-4e20-831d-cf0db0158e81/440b35e6-eecb-4e20-831d-cf0db0158e81.vmdk {{(pid=61243) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1208.720561] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71729638-9aeb-48b0-93da-4022b222abe5 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.742209] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 440b35e6-eecb-4e20-831d-cf0db0158e81/440b35e6-eecb-4e20-831d-cf0db0158e81.vmdk or device None with type streamOptimized {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1208.742494] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b995c6b7-541a-44bc-96c7-6c8403a20dcb {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.761717] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1208.761717] env[61243]: value = "task-1339659" [ 1208.761717] env[61243]: _type = "Task" [ 1208.761717] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.772299] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339659, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.774113] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.637s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.794934] env[61243]: INFO nova.scheduler.client.report [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleted allocations for instance 91b14a22-811a-4151-9769-9c684ec993db [ 1209.271618] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339659, 'name': ReconfigVM_Task, 'duration_secs': 0.331699} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.271905] env[61243]: DEBUG nova.virt.vmwareapi.volumeops [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 440b35e6-eecb-4e20-831d-cf0db0158e81/440b35e6-eecb-4e20-831d-cf0db0158e81.vmdk or device None with type streamOptimized {{(pid=61243) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1209.272546] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-862eee46-0f93-4c2e-a39e-976a0c7e6883 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.278511] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1209.278511] env[61243]: value = "task-1339660" [ 1209.278511] env[61243]: _type = "Task" [ 1209.278511] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.285470] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339660, 'name': Rename_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.301366] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b4306816-166a-4df9-9dbe-d468f225d880 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "91b14a22-811a-4151-9769-9c684ec993db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.530s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.788784] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339660, 'name': Rename_Task} progress is 14%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.289457] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339660, 'name': Rename_Task, 'duration_secs': 0.902546} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.289734] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Powering on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1210.289985] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a68de2d2-d4a4-4bf2-b643-ae920e429e75 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.295832] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1210.295832] env[61243]: value = "task-1339662" [ 1210.295832] env[61243]: _type = "Task" [ 1210.295832] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.302808] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339662, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.806200] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339662, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.308154] env[61243]: DEBUG oslo_vmware.api [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339662, 'name': PowerOnVM_Task, 'duration_secs': 0.511046} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.308388] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Powered on the VM {{(pid=61243) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1211.424222] env[61243]: DEBUG nova.compute.manager [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Checking state {{(pid=61243) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1211.425190] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66eb47e-d4eb-41d6-ba08-d721adaaeeea {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.944433] env[61243]: DEBUG oslo_concurrency.lockutils [None req-b93c9793-3b56-419f-a589-f0e27bac223b tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.293s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.849677] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "440b35e6-eecb-4e20-831d-cf0db0158e81" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.850045] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.850294] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "440b35e6-eecb-4e20-831d-cf0db0158e81-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.850498] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.850677] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.852901] env[61243]: INFO nova.compute.manager [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Terminating instance [ 1212.855051] env[61243]: DEBUG nova.compute.manager [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1212.855262] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1212.856092] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a387203f-467a-4b30-a242-35b270aeefe7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.863459] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1212.863690] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f9ae72b2-826c-49e0-9144-c2d496d8a28e {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.870620] env[61243]: DEBUG oslo_vmware.api [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1212.870620] env[61243]: value = "task-1339663" [ 1212.870620] env[61243]: _type = "Task" [ 1212.870620] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.878638] env[61243]: DEBUG oslo_vmware.api [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339663, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.380412] env[61243]: DEBUG oslo_vmware.api [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339663, 'name': PowerOffVM_Task, 'duration_secs': 0.194582} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.380779] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1213.380779] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1213.381046] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1ceadab9-ed0f-4ec1-8424-8c69edc605da {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.447775] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1213.447998] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1213.448202] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Deleting the datastore file [datastore2] 440b35e6-eecb-4e20-831d-cf0db0158e81 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1213.448668] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d2f69d42-0e6a-4772-87eb-8faf9e676f71 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.454879] env[61243]: DEBUG oslo_vmware.api [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for the task: (returnval){ [ 1213.454879] env[61243]: value = "task-1339665" [ 1213.454879] env[61243]: _type = "Task" [ 1213.454879] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.462335] env[61243]: DEBUG oslo_vmware.api [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339665, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.537289] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "3f839aa5-2b9a-4807-b63b-931f74455532" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1213.537604] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "3f839aa5-2b9a-4807-b63b-931f74455532" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.537878] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "3f839aa5-2b9a-4807-b63b-931f74455532-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1213.538113] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "3f839aa5-2b9a-4807-b63b-931f74455532-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.538307] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "3f839aa5-2b9a-4807-b63b-931f74455532-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.540732] env[61243]: INFO nova.compute.manager [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Terminating instance [ 1213.542759] env[61243]: DEBUG nova.compute.manager [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Start destroying the instance on the hypervisor. {{(pid=61243) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1213.542994] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Destroying instance {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1213.543937] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286d16e6-0aec-4af6-b121-1832653cc0ef {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.551825] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Powering off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1213.552114] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ffd723fd-5a19-4875-8dbe-68e5a4dea081 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.558987] env[61243]: DEBUG oslo_vmware.api [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1213.558987] env[61243]: value = "task-1339666" [ 1213.558987] env[61243]: _type = "Task" [ 1213.558987] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.566733] env[61243]: DEBUG oslo_vmware.api [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339666, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.964769] env[61243]: DEBUG oslo_vmware.api [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Task: {'id': task-1339665, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147053} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.965032] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1213.965233] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1213.965423] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1213.965607] env[61243]: INFO nova.compute.manager [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1213.965854] env[61243]: DEBUG oslo.service.loopingcall [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1213.966063] env[61243]: DEBUG nova.compute.manager [-] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1213.966164] env[61243]: DEBUG nova.network.neutron [-] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1214.070497] env[61243]: DEBUG oslo_vmware.api [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339666, 'name': PowerOffVM_Task, 'duration_secs': 0.189113} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.070822] env[61243]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Powered off the VM {{(pid=61243) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1214.071013] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Unregistering the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1214.071281] env[61243]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1d9fb69-11fd-4add-b8ad-5aa609446ba4 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.130377] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Unregistered the VM {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1214.130585] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Deleting contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1214.130762] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleting the datastore file [datastore2] 3f839aa5-2b9a-4807-b63b-931f74455532 {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1214.131031] env[61243]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7176d5bb-9bba-4b2c-97b3-fef6e5773b9d {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.136980] env[61243]: DEBUG oslo_vmware.api [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for the task: (returnval){ [ 1214.136980] env[61243]: value = "task-1339668" [ 1214.136980] env[61243]: _type = "Task" [ 1214.136980] env[61243]: } to complete. {{(pid=61243) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.144497] env[61243]: DEBUG oslo_vmware.api [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339668, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.403642] env[61243]: DEBUG nova.compute.manager [req-ffbda33d-7d58-42e1-a50a-fce83519aa95 req-f9953ce9-e2db-49e8-bfd8-8f7015343c83 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Received event network-vif-deleted-723d362f-40a3-492a-9a70-f86c847d5d16 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1214.403900] env[61243]: INFO nova.compute.manager [req-ffbda33d-7d58-42e1-a50a-fce83519aa95 req-f9953ce9-e2db-49e8-bfd8-8f7015343c83 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Neutron deleted interface 723d362f-40a3-492a-9a70-f86c847d5d16; detaching it from the instance and deleting it from the info cache [ 1214.404039] env[61243]: DEBUG nova.network.neutron [req-ffbda33d-7d58-42e1-a50a-fce83519aa95 req-f9953ce9-e2db-49e8-bfd8-8f7015343c83 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1214.647150] env[61243]: DEBUG oslo_vmware.api [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Task: {'id': task-1339668, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153077} completed successfully. {{(pid=61243) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.647411] env[61243]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleted the datastore file {{(pid=61243) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1214.647599] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Deleted contents of the VM from datastore datastore2 {{(pid=61243) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1214.647780] env[61243]: DEBUG nova.virt.vmwareapi.vmops [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Instance destroyed {{(pid=61243) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1214.647962] env[61243]: INFO nova.compute.manager [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1214.648224] env[61243]: DEBUG oslo.service.loopingcall [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61243) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1214.648420] env[61243]: DEBUG nova.compute.manager [-] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Deallocating network for instance {{(pid=61243) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1214.648515] env[61243]: DEBUG nova.network.neutron [-] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] deallocate_for_instance() {{(pid=61243) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1214.885533] env[61243]: DEBUG nova.network.neutron [-] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1214.906351] env[61243]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b3a6e45d-4918-4259-814a-c697866bfa92 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.916071] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08bb074d-2770-4519-8cc7-79c233fcf890 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.938278] env[61243]: DEBUG nova.compute.manager [req-ffbda33d-7d58-42e1-a50a-fce83519aa95 req-f9953ce9-e2db-49e8-bfd8-8f7015343c83 service nova] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Detach interface failed, port_id=723d362f-40a3-492a-9a70-f86c847d5d16, reason: Instance 440b35e6-eecb-4e20-831d-cf0db0158e81 could not be found. {{(pid=61243) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1215.368652] env[61243]: DEBUG nova.network.neutron [-] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Updating instance_info_cache with network_info: [] {{(pid=61243) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1215.389839] env[61243]: INFO nova.compute.manager [-] [instance: 440b35e6-eecb-4e20-831d-cf0db0158e81] Took 1.42 seconds to deallocate network for instance. [ 1215.871213] env[61243]: INFO nova.compute.manager [-] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Took 1.22 seconds to deallocate network for instance. [ 1215.895913] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1215.896214] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1215.896440] env[61243]: DEBUG nova.objects.instance [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lazy-loading 'resources' on Instance uuid 440b35e6-eecb-4e20-831d-cf0db0158e81 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1216.377219] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1216.428431] env[61243]: DEBUG nova.compute.manager [req-df08ff55-439b-4885-b93d-dd027188311e req-6cf02237-e06f-4003-bda9-9ae9e71fe8ec service nova] [instance: 3f839aa5-2b9a-4807-b63b-931f74455532] Received event network-vif-deleted-a0b324d6-6a73-4956-85c3-38f755003b74 {{(pid=61243) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1216.440028] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c4d745f-1eac-48b6-b115-adcd7ef29865 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.447991] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0e336b-208f-4c4e-a0c0-6d72de74f7e1 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.477618] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ad7eee-a5f9-463b-a23b-907d239aabe7 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.484625] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749f3a44-0612-430b-9294-a18941b51e71 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.497344] env[61243]: DEBUG nova.compute.provider_tree [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1217.000437] env[61243]: DEBUG nova.scheduler.client.report [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1217.505487] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.609s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.507825] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.131s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1217.508095] env[61243]: DEBUG nova.objects.instance [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lazy-loading 'resources' on Instance uuid 3f839aa5-2b9a-4807-b63b-931f74455532 {{(pid=61243) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1217.525660] env[61243]: INFO nova.scheduler.client.report [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Deleted allocations for instance 440b35e6-eecb-4e20-831d-cf0db0158e81 [ 1218.032338] env[61243]: DEBUG oslo_concurrency.lockutils [None req-6488777a-3a08-43fd-ab55-2f22adcd67ef tempest-ServerActionsTestOtherB-1773685481 tempest-ServerActionsTestOtherB-1773685481-project-member] Lock "440b35e6-eecb-4e20-831d-cf0db0158e81" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.182s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.035863] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcf868b-b365-4c7e-9b17-c5c1bf3e20fa {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.043309] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a370f844-cfd9-4d66-9819-543bbc588732 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.073742] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08fb89cf-9132-4b4b-8411-e31314d6aa05 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.080732] env[61243]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e85b0a9-f0ce-4cfa-8764-20d0fb2c9208 {{(pid=61243) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.093622] env[61243]: DEBUG nova.compute.provider_tree [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed in ProviderTree for provider: 8f3c282f-58fe-4c5d-80db-5a142cf023e1 {{(pid=61243) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1218.597027] env[61243]: DEBUG nova.scheduler.client.report [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Inventory has not changed for provider 8f3c282f-58fe-4c5d-80db-5a142cf023e1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61243) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1219.102067] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.594s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1219.117718] env[61243]: INFO nova.scheduler.client.report [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Deleted allocations for instance 3f839aa5-2b9a-4807-b63b-931f74455532 [ 1219.625303] env[61243]: DEBUG oslo_concurrency.lockutils [None req-f6d3cf55-9b00-4d74-b254-1897664d9bd2 tempest-ServerActionsTestOtherA-385971820 tempest-ServerActionsTestOtherA-385971820-project-member] Lock "3f839aa5-2b9a-4807-b63b-931f74455532" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.088s {{(pid=61243) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}